diff --git a/.gitattributes b/.gitattributes index b3d487871e5124acc2b73a0246a935b6cc85c74b..217cb4d692349619a6835f70d937b350311a92f9 100644 --- a/.gitattributes +++ b/.gitattributes @@ -173,3 +173,14 @@ runs/l2r50-baseline/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs runs/l2r50-baseline/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text runs/l2r50-baseline/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text runs/l2r50-baseline/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-1080/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-1440/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-2160/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-2520/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-2880/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-3240/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r90-baseline/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/runs/l2r90-baseline/checkpoint-1080/chat_template.jinja b/runs/l2r90-baseline/checkpoint-1080/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-1080/config.json b/runs/l2r90-baseline/checkpoint-1080/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-1080/generation_config.json b/runs/l2r90-baseline/checkpoint-1080/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-1080/model.safetensors b/runs/l2r90-baseline/checkpoint-1080/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6db7d84c02f1287ee824c056b6fd8e0a224beb4a --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c913a6ecc6b022c054063e727c029a2bff92dcf3208f8d9ba5977ff69619f497 +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-1080/optimizer.pt b/runs/l2r90-baseline/checkpoint-1080/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..d7c5a6c8c931aed33982d6c49c6e3aee526befe5 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ebe66f5632d12006244fc83fa147e0ced7b1ca7f46ba26359253c46df20f8b1 +size 1166825338 diff --git a/runs/l2r90-baseline/checkpoint-1080/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-1080/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..691b2ea3e1b4f187dd5e6c0cda81e8858f6ac4cf --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25e1b415517c5fe22089b6e5af2651408c03277402e7a2e6910a756d01bc912d +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1080/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-1080/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..8832e70f2d40e03d52900feac4951dac7fa3ec89 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5eac241b5f2b35e77529d15f1fe989e6cbe8bfc9e18a3b3e5d842da1f82ab4e2 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1080/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-1080/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..cc6afe867703008555530c3c86fd6cf35ab7d7b1 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2e72c528a5f7c6676fb4c9abd29d59ee303309497199e2b8bea743dbd9d88db +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1080/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-1080/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..4eb9917247f226d9cb7ed167e78dae92754ce06b --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40d0c6d939e35fba984de75edb8ae6eeab9a914372a646423e26bfff4aff7f03 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1080/scheduler.pt b/runs/l2r90-baseline/checkpoint-1080/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..6460bf7d5e8aadff08654c5c446813d18e721d62 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd8f1c3c3cf1427d3265ecb044cf76510643402bb89ce35df8f605375b066d14 +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-1080/tokenizer.json b/runs/l2r90-baseline/checkpoint-1080/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-1080/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-1080/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-1080/trainer_state.json b/runs/l2r90-baseline/checkpoint-1080/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..f067db6195192ed99b6fbd090569af5aaa385d7f --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/trainer_state.json @@ -0,0 +1,7706 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 15.0, + "eval_steps": 500, + "global_step": 1080, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 5.549652099609375, + "eval_runtime": 57.761, + "eval_samples_per_second": 42.278, + "eval_steps_per_second": 1.333, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.03958690166473389, + "learning_rate": 0.0005919487226162231, + "loss": 5.459286689758301, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.03299890458583832, + "learning_rate": 0.0005918869546699877, + "loss": 5.45464563369751, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.028040863573551178, + "learning_rate": 0.0005918249539727509, + "loss": 5.444796562194824, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.028365720063447952, + "learning_rate": 0.0005917627205744655, + "loss": 5.447852611541748, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.02822224237024784, + "learning_rate": 0.0005917002545252719, + "loss": 5.453883171081543, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.036075398325920105, + "learning_rate": 0.0005916375558754977, + "loss": 5.462671279907227, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04683876410126686, + "learning_rate": 0.0005915746246756581, + "loss": 5.443917274475098, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.05457588657736778, + "learning_rate": 0.0005915114609764558, + "loss": 5.444374084472656, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.04250836744904518, + "learning_rate": 0.0005914480648287804, + "loss": 5.435029983520508, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.029477844014763832, + "learning_rate": 0.0005913844362837093, + "loss": 5.421862602233887, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.03673284500837326, + "learning_rate": 0.0005913205753925066, + "loss": 5.430391311645508, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.030489861965179443, + "learning_rate": 0.0005912564822066241, + "loss": 5.389314651489258, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.04037046805024147, + "learning_rate": 0.0005911921567777005, + "loss": 5.425060272216797, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.0445217490196228, + "learning_rate": 0.0005911275991575613, + "loss": 5.437565326690674, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.03885014355182648, + "learning_rate": 0.0005910628093982198, + "loss": 5.410930633544922, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.0334964245557785, + "learning_rate": 0.0005909977875518759, + "loss": 5.412092208862305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.03499286621809006, + "learning_rate": 0.0005909325336709164, + "loss": 5.382216453552246, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.04171644523739815, + "learning_rate": 0.0005908670478079152, + "loss": 5.387385368347168, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.057558078318834305, + "learning_rate": 0.000590801330015633, + "loss": 5.437925815582275, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.053846780210733414, + "learning_rate": 0.0005907353803470177, + "loss": 5.41644811630249, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.027259036898612976, + "learning_rate": 0.0005906691988552034, + "loss": 5.409185409545898, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.03394043818116188, + "learning_rate": 0.0005906027855935115, + "loss": 5.377973556518555, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.03731574863195419, + "learning_rate": 0.0005905361406154501, + "loss": 5.382505893707275, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.04705141484737396, + "learning_rate": 0.0005904692639747137, + "loss": 5.413661003112793, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.044811323285102844, + "learning_rate": 0.0005904021557251837, + "loss": 5.417835235595703, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.04000752419233322, + "learning_rate": 0.0005903348159209277, + "loss": 5.406160354614258, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.038364164531230927, + "learning_rate": 0.0005902672446162007, + "loss": 5.357950687408447, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.030500127002596855, + "learning_rate": 0.0005901994418654432, + "loss": 5.344951629638672, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.030644729733467102, + "learning_rate": 0.0005901314077232829, + "loss": 5.3599348068237305, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.037976790219545364, + "learning_rate": 0.0005900631422445335, + "loss": 5.354084491729736, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.04106011986732483, + "learning_rate": 0.0005899946454841955, + "loss": 5.363678455352783, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.04334832355380058, + "learning_rate": 0.0005899259174974553, + "loss": 5.350959300994873, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.049939028918743134, + "learning_rate": 0.000589856958339686, + "loss": 5.3317155838012695, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.05300058051943779, + "learning_rate": 0.0005897877680664465, + "loss": 5.353314399719238, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.05943184345960617, + "learning_rate": 0.0005897183467334822, + "loss": 5.3683929443359375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.05600789934396744, + "learning_rate": 0.0005896486943967249, + "loss": 5.345516204833984, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.05392718315124512, + "learning_rate": 0.0005895788111122919, + "loss": 5.291389465332031, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.037913475185632706, + "learning_rate": 0.0005895086969364871, + "loss": 5.353423118591309, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.03362123295664787, + "learning_rate": 0.0005894383519258001, + "loss": 5.379500389099121, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032558735460042953, + "learning_rate": 0.0005893677761369066, + "loss": 5.34043025970459, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.04074402526021004, + "learning_rate": 0.0005892969696266683, + "loss": 5.303869247436523, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.04270484298467636, + "learning_rate": 0.0005892259324521328, + "loss": 5.27297830581665, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.04604959115386009, + "learning_rate": 0.0005891546646705334, + "loss": 5.313268184661865, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.04462246224284172, + "learning_rate": 0.0005890831663392891, + "loss": 5.31167459487915, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.04966094344854355, + "learning_rate": 0.0005890114375160052, + "loss": 5.316695690155029, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.04662073776125908, + "learning_rate": 0.0005889394782584718, + "loss": 5.316898822784424, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.04380315542221069, + "learning_rate": 0.0005888672886246656, + "loss": 5.285087585449219, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03239579498767853, + "learning_rate": 0.0005887948686727483, + "loss": 5.304283142089844, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.029690183699131012, + "learning_rate": 0.0005887222184610675, + "loss": 5.305715084075928, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.03444967418909073, + "learning_rate": 0.0005886493380481559, + "loss": 5.2707672119140625, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.03302565962076187, + "learning_rate": 0.0005885762274927322, + "loss": 5.277251243591309, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.028477763757109642, + "learning_rate": 0.0005885028868537, + "loss": 5.295318603515625, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.022988801822066307, + "learning_rate": 0.0005884293161901487, + "loss": 5.249057292938232, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.026115216314792633, + "learning_rate": 0.000588355515561353, + "loss": 5.270636081695557, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.02576332725584507, + "learning_rate": 0.0005882814850267723, + "loss": 5.263096332550049, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.023978352546691895, + "learning_rate": 0.0005882072246460521, + "loss": 5.276605606079102, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.020695513114333153, + "learning_rate": 0.0005881327344790223, + "loss": 5.2141432762146, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01952219195663929, + "learning_rate": 0.0005880580145856986, + "loss": 5.214289665222168, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.022633206099271774, + "learning_rate": 0.0005879830650262813, + "loss": 5.236690998077393, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.023082347586750984, + "learning_rate": 0.0005879078858611557, + "loss": 5.2092156410217285, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.019620463252067566, + "learning_rate": 0.0005878324771508927, + "loss": 5.264451503753662, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.021504493430256844, + "learning_rate": 0.0005877568389562475, + "loss": 5.23444938659668, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.025672364979982376, + "learning_rate": 0.0005876809713381606, + "loss": 5.205289363861084, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.034436631947755814, + "learning_rate": 0.0005876048743577569, + "loss": 5.2047648429870605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.042389530688524246, + "learning_rate": 0.0005875285480763466, + "loss": 5.192008018493652, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.03310716897249222, + "learning_rate": 0.0005874519925554244, + "loss": 5.162734031677246, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.03029288910329342, + "learning_rate": 0.0005873752078566694, + "loss": 5.212151527404785, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04619310051202774, + "learning_rate": 0.0005872981940419459, + "loss": 5.215909004211426, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.05618087574839592, + "learning_rate": 0.0005872209511733027, + "loss": 5.210192680358887, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.03783707693219185, + "learning_rate": 0.0005871434793129726, + "loss": 5.201597690582275, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.03721104562282562, + "learning_rate": 0.0005870657785233735, + "loss": 5.188891410827637, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.033336155116558075, + "learning_rate": 0.0005869878488671075, + "loss": 5.1598615646362305, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.2592363357543945, + "eval_runtime": 57.3343, + "eval_samples_per_second": 42.592, + "eval_steps_per_second": 1.343, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03719741478562355, + "learning_rate": 0.0005869096904069611, + "loss": 5.1646623611450195, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.04888090118765831, + "learning_rate": 0.0005868313032059052, + "loss": 5.171182155609131, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.05073149502277374, + "learning_rate": 0.0005867526873270949, + "loss": 5.183949947357178, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.04187731072306633, + "learning_rate": 0.0005866738428338695, + "loss": 5.176133155822754, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.039445552974939346, + "learning_rate": 0.000586594769789753, + "loss": 5.148100852966309, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.035551708191633224, + "learning_rate": 0.0005865154682584524, + "loss": 5.151512145996094, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.031175289303064346, + "learning_rate": 0.0005864359383038602, + "loss": 5.102468967437744, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.03696789592504501, + "learning_rate": 0.000586356179990052, + "loss": 5.119902610778809, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.04836135357618332, + "learning_rate": 0.0005862761933812875, + "loss": 5.15183162689209, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.05206914246082306, + "learning_rate": 0.0005861959785420106, + "loss": 5.154870986938477, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.046368084847927094, + "learning_rate": 0.000586115535536849, + "loss": 5.130651473999023, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.07146432995796204, + "learning_rate": 0.000586034864430614, + "loss": 5.175804138183594, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.08809789270162582, + "learning_rate": 0.0005859539652883008, + "loss": 5.229716777801514, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.08116360008716583, + "learning_rate": 0.0005858728381750888, + "loss": 5.21910285949707, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.08282570540904999, + "learning_rate": 0.0005857914831563402, + "loss": 5.2336506843566895, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.07403547316789627, + "learning_rate": 0.0005857099002976016, + "loss": 5.251604080200195, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.07024004310369492, + "learning_rate": 0.0005856280896646026, + "loss": 5.267114162445068, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.056330062448978424, + "learning_rate": 0.0005855460513232567, + "loss": 5.201976299285889, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.051763955503702164, + "learning_rate": 0.0005854637853396606, + "loss": 5.1733198165893555, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.039900992065668106, + "learning_rate": 0.0005853812917800945, + "loss": 5.222941875457764, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.04187863692641258, + "learning_rate": 0.0005852985707110221, + "loss": 5.165033340454102, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.042645812034606934, + "learning_rate": 0.0005852156221990901, + "loss": 5.157554626464844, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.03856075555086136, + "learning_rate": 0.0005851324463111289, + "loss": 5.184629440307617, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.029387330636382103, + "learning_rate": 0.0005850490431141516, + "loss": 5.118739128112793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.027584781870245934, + "learning_rate": 0.0005849654126753545, + "loss": 5.100855350494385, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.023655924946069717, + "learning_rate": 0.0005848815550621175, + "loss": 5.154520034790039, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.024163050577044487, + "learning_rate": 0.0005847974703420028, + "loss": 5.105189323425293, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.020369920879602432, + "learning_rate": 0.0005847131585827564, + "loss": 5.113671779632568, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.021030737087130547, + "learning_rate": 0.0005846286198523061, + "loss": 5.0847649574279785, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.01729249767959118, + "learning_rate": 0.0005845438542187638, + "loss": 5.053225517272949, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.018704712390899658, + "learning_rate": 0.0005844588617504236, + "loss": 5.079106330871582, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.02049064077436924, + "learning_rate": 0.0005843736425157621, + "loss": 5.044185161590576, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.020736856386065483, + "learning_rate": 0.000584288196583439, + "loss": 5.041932106018066, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.016543276607990265, + "learning_rate": 0.0005842025240222966, + "loss": 5.032522201538086, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.018414106220006943, + "learning_rate": 0.0005841166249013598, + "loss": 5.037397384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.01611698977649212, + "learning_rate": 0.0005840304992898359, + "loss": 5.038179397583008, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.014310561120510101, + "learning_rate": 0.0005839441472571147, + "loss": 5.0146484375, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.015472214668989182, + "learning_rate": 0.0005838575688727685, + "loss": 5.060323715209961, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.016373401507735252, + "learning_rate": 0.0005837707642065518, + "loss": 5.032100677490234, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.018694665282964706, + "learning_rate": 0.000583683733328402, + "loss": 5.036993026733398, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.019977087154984474, + "learning_rate": 0.0005835964763084378, + "loss": 4.991037368774414, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.023741086944937706, + "learning_rate": 0.000583508993216961, + "loss": 5.043832778930664, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.025061562657356262, + "learning_rate": 0.0005834212841244548, + "loss": 5.016516208648682, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.020848384127020836, + "learning_rate": 0.0005833333491015851, + "loss": 4.986289978027344, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.015086354687809944, + "learning_rate": 0.0005832451882191995, + "loss": 4.9948625564575195, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.015096964314579964, + "learning_rate": 0.0005831568015483274, + "loss": 4.996342658996582, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016932200640439987, + "learning_rate": 0.0005830681891601807, + "loss": 4.998758316040039, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.019936930388212204, + "learning_rate": 0.0005829793511261525, + "loss": 4.959854602813721, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.027658682316541672, + "learning_rate": 0.0005828902875178183, + "loss": 4.968629837036133, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.03299275413155556, + "learning_rate": 0.0005828009984069347, + "loss": 4.994814872741699, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.033717650920152664, + "learning_rate": 0.0005827114838654406, + "loss": 4.976295471191406, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.02969924360513687, + "learning_rate": 0.0005826217439654562, + "loss": 4.969155788421631, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.02846827358007431, + "learning_rate": 0.0005825317787792831, + "loss": 4.989721298217773, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.03670600801706314, + "learning_rate": 0.0005824415883794049, + "loss": 4.988689422607422, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.045570675283670425, + "learning_rate": 0.0005823511728384863, + "loss": 4.974407196044922, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.04564111679792404, + "learning_rate": 0.0005822605322293733, + "loss": 4.966331481933594, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.04173210263252258, + "learning_rate": 0.0005821696666250937, + "loss": 4.925197601318359, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.04228242486715317, + "learning_rate": 0.0005820785760988559, + "loss": 4.965020656585693, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.05144782364368439, + "learning_rate": 0.0005819872607240503, + "loss": 5.0022149085998535, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.049400657415390015, + "learning_rate": 0.0005818957205742478, + "loss": 4.9635114669799805, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.043692708015441895, + "learning_rate": 0.0005818039557232005, + "loss": 4.977723121643066, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.04891742020845413, + "learning_rate": 0.0005817119662448421, + "loss": 4.996312141418457, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.035564981400966644, + "learning_rate": 0.0005816197522132866, + "loss": 4.940932273864746, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04679827764630318, + "learning_rate": 0.0005815273137028292, + "loss": 4.983131408691406, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.05489890277385712, + "learning_rate": 0.0005814346507879459, + "loss": 4.977827548980713, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04640809819102287, + "learning_rate": 0.0005813417635432937, + "loss": 4.954761505126953, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04439452290534973, + "learning_rate": 0.00058124865204371, + "loss": 4.914865970611572, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.04587096720933914, + "learning_rate": 0.0005811553163642131, + "loss": 4.949592590332031, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.044119734317064285, + "learning_rate": 0.0005810617565800019, + "loss": 4.956068992614746, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.043133001774549484, + "learning_rate": 0.0005809679727664559, + "loss": 4.993522644042969, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04315278306603432, + "learning_rate": 0.0005808739649991348, + "loss": 4.894291877746582, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04819472134113312, + "learning_rate": 0.0005807797333537792, + "loss": 4.958306789398193, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.023706436157227, + "eval_runtime": 57.4082, + "eval_samples_per_second": 42.537, + "eval_steps_per_second": 1.341, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.04403403028845787, + "learning_rate": 0.0005806852779063098, + "loss": 4.884971618652344, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.041993748396635056, + "learning_rate": 0.0005805905987328275, + "loss": 4.916353225708008, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.042302004992961884, + "learning_rate": 0.0005804956959096138, + "loss": 4.9012556076049805, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.044500596821308136, + "learning_rate": 0.0005804005695131299, + "loss": 4.915240287780762, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.03914736211299896, + "learning_rate": 0.0005803052196200177, + "loss": 4.885915756225586, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.0361703559756279, + "learning_rate": 0.0005802096463070988, + "loss": 4.829287528991699, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03432302549481392, + "learning_rate": 0.0005801138496513749, + "loss": 4.926834583282471, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.0317804180085659, + "learning_rate": 0.0005800178297300276, + "loss": 4.8707170486450195, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.035261351615190506, + "learning_rate": 0.0005799215866204183, + "loss": 4.897791862487793, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.03276067599654198, + "learning_rate": 0.0005798251204000886, + "loss": 4.864504814147949, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.034757066518068314, + "learning_rate": 0.0005797284311467594, + "loss": 4.890812873840332, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.033726081252098083, + "learning_rate": 0.0005796315189383316, + "loss": 4.831684112548828, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.03074726276099682, + "learning_rate": 0.0005795343838528855, + "loss": 4.8534111976623535, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02215893566608429, + "learning_rate": 0.0005794370259686811, + "loss": 4.8311238288879395, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.02132033370435238, + "learning_rate": 0.0005793394453641581, + "loss": 4.842617034912109, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020825816318392754, + "learning_rate": 0.0005792416421179351, + "loss": 4.84535026550293, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.019334357231855392, + "learning_rate": 0.0005791436163088106, + "loss": 4.839937686920166, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017372334375977516, + "learning_rate": 0.0005790453680157622, + "loss": 4.8115553855896, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.0190406683832407, + "learning_rate": 0.0005789468973179468, + "loss": 4.832542419433594, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.017372574657201767, + "learning_rate": 0.0005788482042947004, + "loss": 4.793632507324219, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.019171632826328278, + "learning_rate": 0.0005787492890255382, + "loss": 4.776987552642822, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.02109278179705143, + "learning_rate": 0.0005786501515901545, + "loss": 4.8216657638549805, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.026234494522213936, + "learning_rate": 0.0005785507920684225, + "loss": 4.8109049797058105, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.0357641875743866, + "learning_rate": 0.0005784512105403944, + "loss": 4.821109771728516, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.03627947345376015, + "learning_rate": 0.000578351407086301, + "loss": 4.810632228851318, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.032253146171569824, + "learning_rate": 0.0005782513817865527, + "loss": 4.7930521965026855, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.03402210772037506, + "learning_rate": 0.0005781511347217375, + "loss": 4.772500991821289, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.0321584977209568, + "learning_rate": 0.000578050665972623, + "loss": 4.781399726867676, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.030242707580327988, + "learning_rate": 0.0005779499756201549, + "loss": 4.802542686462402, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.02361336164176464, + "learning_rate": 0.0005778490637454576, + "loss": 4.771912574768066, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.028781302273273468, + "learning_rate": 0.000577747930429834, + "loss": 4.796924591064453, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.027042683213949203, + "learning_rate": 0.0005776465757547654, + "loss": 4.745218276977539, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.02715984545648098, + "learning_rate": 0.0005775449998019113, + "loss": 4.792944431304932, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.0314096100628376, + "learning_rate": 0.00057744320265311, + "loss": 4.7649993896484375, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.028749193996191025, + "learning_rate": 0.0005773411843903773, + "loss": 4.7980546951293945, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.02655309997498989, + "learning_rate": 0.0005772389450959075, + "loss": 4.746520519256592, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.027497397735714912, + "learning_rate": 0.000577136484852073, + "loss": 4.7490715980529785, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.0343393050134182, + "learning_rate": 0.000577033803741424, + "loss": 4.74436092376709, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.04444412514567375, + "learning_rate": 0.0005769309018466891, + "loss": 4.755973815917969, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.0437556728720665, + "learning_rate": 0.0005768277792507744, + "loss": 4.784368515014648, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0369921550154686, + "learning_rate": 0.0005767244360367638, + "loss": 4.753410339355469, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.05038141459226608, + "learning_rate": 0.0005766208722879192, + "loss": 4.755491256713867, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.04511422663927078, + "learning_rate": 0.00057651708808768, + "loss": 4.760293006896973, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.048762883991003036, + "learning_rate": 0.0005764130835196632, + "loss": 4.755958080291748, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.06749686598777771, + "learning_rate": 0.0005763088586676634, + "loss": 4.802557468414307, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.06451250612735748, + "learning_rate": 0.0005762044136156528, + "loss": 4.812925338745117, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.05436156690120697, + "learning_rate": 0.0005760997484477809, + "loss": 4.807103157043457, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.05233754590153694, + "learning_rate": 0.0005759948632483742, + "loss": 4.782449722290039, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.05519746616482735, + "learning_rate": 0.0005758897581019371, + "loss": 4.808449745178223, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.0578567236661911, + "learning_rate": 0.000575784433093151, + "loss": 4.7676005363464355, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.0588264986872673, + "learning_rate": 0.0005756788883068743, + "loss": 4.805377960205078, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04225790873169899, + "learning_rate": 0.0005755731238281423, + "loss": 4.748775005340576, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.0376368910074234, + "learning_rate": 0.0005754671397421678, + "loss": 4.759791374206543, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03830771893262863, + "learning_rate": 0.0005753609361343402, + "loss": 4.729974746704102, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.03684520348906517, + "learning_rate": 0.0005752545130902256, + "loss": 4.729854583740234, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02983408235013485, + "learning_rate": 0.0005751478706955674, + "loss": 4.78502082824707, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02854904904961586, + "learning_rate": 0.0005750410090362854, + "loss": 4.717472076416016, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.023819150403141975, + "learning_rate": 0.0005749339281984761, + "loss": 4.717124938964844, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.02477158047258854, + "learning_rate": 0.0005748266282684124, + "loss": 4.71781063079834, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.022002534940838814, + "learning_rate": 0.0005747191093325443, + "loss": 4.7480669021606445, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.02133483812212944, + "learning_rate": 0.0005746113714774976, + "loss": 4.693770408630371, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.022407138720154762, + "learning_rate": 0.0005745034147900747, + "loss": 4.720734119415283, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.020850926637649536, + "learning_rate": 0.0005743952393572544, + "loss": 4.745505332946777, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.020420582965016365, + "learning_rate": 0.0005742868452661918, + "loss": 4.688610076904297, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.019014107063412666, + "learning_rate": 0.0005741782326042181, + "loss": 4.673436164855957, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.017393039539456367, + "learning_rate": 0.0005740694014588403, + "loss": 4.6941118240356445, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.017877329140901566, + "learning_rate": 0.0005739603519177419, + "loss": 4.667474269866943, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.017749635502696037, + "learning_rate": 0.0005738510840687821, + "loss": 4.6901092529296875, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.018746376037597656, + "learning_rate": 0.000573741597999996, + "loss": 4.680166721343994, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016260217875242233, + "learning_rate": 0.0005736318937995947, + "loss": 4.6799821853637695, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015099814161658287, + "learning_rate": 0.0005735219715559646, + "loss": 4.680666923522949, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.016539735719561577, + "learning_rate": 0.0005734118313576683, + "loss": 4.639509201049805, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 4.768754482269287, + "eval_runtime": 58.3352, + "eval_samples_per_second": 41.862, + "eval_steps_per_second": 1.32, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.019733518362045288, + "learning_rate": 0.0005733014732934436, + "loss": 4.635312080383301, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.022571036592125893, + "learning_rate": 0.0005731908974522042, + "loss": 4.6041765213012695, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.022238552570343018, + "learning_rate": 0.0005730801039230389, + "loss": 4.648976802825928, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.02115180902183056, + "learning_rate": 0.0005729690927952123, + "loss": 4.623968124389648, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.021789928898215294, + "learning_rate": 0.0005728578641581637, + "loss": 4.620028018951416, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.02593887411057949, + "learning_rate": 0.0005727464181015081, + "loss": 4.591062545776367, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.0347842313349247, + "learning_rate": 0.0005726347547150357, + "loss": 4.632149696350098, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04119519516825676, + "learning_rate": 0.0005725228740887117, + "loss": 4.6254191398620605, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.04895750433206558, + "learning_rate": 0.0005724107763126761, + "loss": 4.621372222900391, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.05752161890268326, + "learning_rate": 0.0005722984614772442, + "loss": 4.61965274810791, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05115870013833046, + "learning_rate": 0.000572185929672906, + "loss": 4.63119649887085, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.05721089243888855, + "learning_rate": 0.0005720731809903263, + "loss": 4.652647018432617, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.0646839588880539, + "learning_rate": 0.000571960215520345, + "loss": 4.642184257507324, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.0682854950428009, + "learning_rate": 0.0005718470333539757, + "loss": 4.676589012145996, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.057402849197387695, + "learning_rate": 0.0005717336345824077, + "loss": 4.6585283279418945, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.05885373800992966, + "learning_rate": 0.0005716200192970043, + "loss": 4.671728134155273, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.05265705659985542, + "learning_rate": 0.0005715061875893032, + "loss": 4.641249656677246, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.05417842045426369, + "learning_rate": 0.0005713921395510166, + "loss": 4.661381721496582, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.06602369248867035, + "learning_rate": 0.0005712778752740307, + "loss": 4.684137344360352, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.06004182621836662, + "learning_rate": 0.0005711633948504066, + "loss": 4.625332832336426, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.04801511764526367, + "learning_rate": 0.0005710486983723787, + "loss": 4.664791107177734, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.04380105435848236, + "learning_rate": 0.0005709337859323561, + "loss": 4.664966583251953, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.03858727216720581, + "learning_rate": 0.0005708186576229218, + "loss": 4.661867618560791, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03240576758980751, + "learning_rate": 0.0005707033135368323, + "loss": 4.5855326652526855, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.028561925515532494, + "learning_rate": 0.0005705877537670184, + "loss": 4.637331008911133, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.026389246806502342, + "learning_rate": 0.0005704719784065846, + "loss": 4.598666667938232, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.02484819106757641, + "learning_rate": 0.0005703559875488088, + "loss": 4.601729393005371, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.02522977627813816, + "learning_rate": 0.0005702397812871429, + "loss": 4.609833717346191, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02589508146047592, + "learning_rate": 0.0005701233597152121, + "loss": 4.574592590332031, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.023374728858470917, + "learning_rate": 0.0005700067229268154, + "loss": 4.61362361907959, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.02465851791203022, + "learning_rate": 0.0005698898710159245, + "loss": 4.639899253845215, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.02258715033531189, + "learning_rate": 0.0005697728040766852, + "loss": 4.5656538009643555, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.018009789288043976, + "learning_rate": 0.0005696555222034162, + "loss": 4.61449670791626, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02121281623840332, + "learning_rate": 0.0005695380254906094, + "loss": 4.553830623626709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.018078701570630074, + "learning_rate": 0.0005694203140329296, + "loss": 4.534579277038574, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.0169303547590971, + "learning_rate": 0.000569302387925215, + "loss": 4.533995628356934, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.016318142414093018, + "learning_rate": 0.0005691842472624764, + "loss": 4.544438362121582, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.016664575785398483, + "learning_rate": 0.0005690658921398977, + "loss": 4.5913286209106445, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.01641504280269146, + "learning_rate": 0.0005689473226528354, + "loss": 4.520743370056152, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.017012735828757286, + "learning_rate": 0.0005688285388968187, + "loss": 4.517453193664551, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01672312617301941, + "learning_rate": 0.0005687095409675499, + "loss": 4.5235915184021, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.016893865540623665, + "learning_rate": 0.000568590328960903, + "loss": 4.543200492858887, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.017530135810375214, + "learning_rate": 0.0005684709029729253, + "loss": 4.54912805557251, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.018698599189519882, + "learning_rate": 0.0005683512630998361, + "loss": 4.522650718688965, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01984071172773838, + "learning_rate": 0.000568231409438027, + "loss": 4.546173095703125, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.023218289017677307, + "learning_rate": 0.0005681113420840619, + "loss": 4.5305094718933105, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.028968270868062973, + "learning_rate": 0.000567991061134677, + "loss": 4.521440505981445, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.032370273023843765, + "learning_rate": 0.0005678705666867805, + "loss": 4.54168701171875, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.033967748284339905, + "learning_rate": 0.0005677498588374524, + "loss": 4.538717269897461, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.028757929801940918, + "learning_rate": 0.0005676289376839452, + "loss": 4.517228603363037, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.029757825657725334, + "learning_rate": 0.0005675078033236827, + "loss": 4.510310173034668, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.03581003099679947, + "learning_rate": 0.0005673864558542605, + "loss": 4.506901264190674, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.03483985736966133, + "learning_rate": 0.0005672648953734462, + "loss": 4.549284934997559, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.031016027554869652, + "learning_rate": 0.000567143121979179, + "loss": 4.530084609985352, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.029373526573181152, + "learning_rate": 0.0005670211357695693, + "loss": 4.492218971252441, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.031661372631788254, + "learning_rate": 0.0005668989368428994, + "loss": 4.502824783325195, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.027251387014985085, + "learning_rate": 0.0005667765252976227, + "loss": 4.49193811416626, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02263176068663597, + "learning_rate": 0.0005666539012323639, + "loss": 4.512019157409668, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.02355228364467621, + "learning_rate": 0.0005665310647459189, + "loss": 4.508917331695557, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.023297371342778206, + "learning_rate": 0.0005664080159372551, + "loss": 4.497771263122559, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.019364114850759506, + "learning_rate": 0.0005662847549055107, + "loss": 4.51470947265625, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.023899901658296585, + "learning_rate": 0.0005661612817499947, + "loss": 4.48771858215332, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.02330983616411686, + "learning_rate": 0.0005660375965701872, + "loss": 4.478496551513672, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.027252234518527985, + "learning_rate": 0.0005659136994657392, + "loss": 4.497672080993652, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.030886363238096237, + "learning_rate": 0.0005657895905364723, + "loss": 4.49789571762085, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02515559457242489, + "learning_rate": 0.0005656652698823788, + "loss": 4.490771293640137, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020107125863432884, + "learning_rate": 0.0005655407376036217, + "loss": 4.481879234313965, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.022076716646552086, + "learning_rate": 0.0005654159938005345, + "loss": 4.494697570800781, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.025525815784931183, + "learning_rate": 0.0005652910385736207, + "loss": 4.517453193664551, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.03005894646048546, + "learning_rate": 0.0005651658720235548, + "loss": 4.453815460205078, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.030178502202033997, + "learning_rate": 0.0005650404942511812, + "loss": 4.493922233581543, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.029249394312500954, + "learning_rate": 0.0005649149053575141, + "loss": 4.453763008117676, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 4.607339382171631, + "eval_runtime": 57.9216, + "eval_samples_per_second": 42.16, + "eval_steps_per_second": 1.329, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.03362950682640076, + "learning_rate": 0.0005647891054437386, + "loss": 4.430700302124023, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.04223684221506119, + "learning_rate": 0.0005646630946112096, + "loss": 4.486915588378906, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.05192171037197113, + "learning_rate": 0.0005645368729614512, + "loss": 4.467541694641113, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.042525868862867355, + "learning_rate": 0.0005644104405961582, + "loss": 4.45252799987793, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.0406370609998703, + "learning_rate": 0.0005642837976171949, + "loss": 4.414548873901367, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.044476233422756195, + "learning_rate": 0.0005641569441265952, + "loss": 4.453492164611816, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.04687326401472092, + "learning_rate": 0.0005640298802265626, + "loss": 4.477725028991699, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.04011611267924309, + "learning_rate": 0.0005639026060194704, + "loss": 4.423356533050537, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.04971884936094284, + "learning_rate": 0.0005637751216078609, + "loss": 4.4853515625, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.05863234028220177, + "learning_rate": 0.0005636474270944461, + "loss": 4.466001033782959, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.057710036635398865, + "learning_rate": 0.0005635195225821072, + "loss": 4.466723442077637, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.05155074968934059, + "learning_rate": 0.0005633914081738945, + "loss": 4.4823198318481445, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.04535634070634842, + "learning_rate": 0.0005632630839730275, + "loss": 4.471219062805176, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.045147854834795, + "learning_rate": 0.0005631345500828946, + "loss": 4.476888179779053, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.03730619326233864, + "learning_rate": 0.0005630058066070533, + "loss": 4.428932189941406, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.034599728882312775, + "learning_rate": 0.0005628768536492299, + "loss": 4.47627592086792, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.032904356718063354, + "learning_rate": 0.0005627476913133193, + "loss": 4.422671794891357, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.03356337174773216, + "learning_rate": 0.0005626183197033856, + "loss": 4.456406593322754, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.035042595118284225, + "learning_rate": 0.0005624887389236609, + "loss": 4.371820449829102, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.03269032761454582, + "learning_rate": 0.0005623589490785462, + "loss": 4.402605056762695, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.028261031955480576, + "learning_rate": 0.0005622289502726108, + "loss": 4.433136940002441, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03301700949668884, + "learning_rate": 0.0005620987426105925, + "loss": 4.388359546661377, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.031664177775382996, + "learning_rate": 0.000561968326197397, + "loss": 4.41859245300293, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.026281699538230896, + "learning_rate": 0.0005618377011380988, + "loss": 4.406128883361816, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.023097911849617958, + "learning_rate": 0.0005617068675379398, + "loss": 4.389717102050781, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.022349286824464798, + "learning_rate": 0.0005615758255023306, + "loss": 4.408173561096191, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.021958226338028908, + "learning_rate": 0.0005614445751368491, + "loss": 4.434830188751221, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024551713839173317, + "learning_rate": 0.0005613131165472415, + "loss": 4.386944770812988, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.028048967942595482, + "learning_rate": 0.0005611814498394216, + "loss": 4.402279376983643, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02733907848596573, + "learning_rate": 0.0005610495751194708, + "loss": 4.403069496154785, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.020858559757471085, + "learning_rate": 0.0005609174924936384, + "loss": 4.406920909881592, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.021491149440407753, + "learning_rate": 0.0005607852020683409, + "loss": 4.385799407958984, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.021495329216122627, + "learning_rate": 0.0005606527039501621, + "loss": 4.4252495765686035, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.019708355888724327, + "learning_rate": 0.0005605199982458535, + "loss": 4.388864994049072, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.01886274851858616, + "learning_rate": 0.0005603870850623338, + "loss": 4.335962295532227, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.01955341547727585, + "learning_rate": 0.0005602539645066884, + "loss": 4.373560905456543, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.018962012603878975, + "learning_rate": 0.0005601206366861706, + "loss": 4.380022048950195, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.016663335263729095, + "learning_rate": 0.0005599871017081999, + "loss": 4.421034336090088, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01676873490214348, + "learning_rate": 0.0005598533596803631, + "loss": 4.382775783538818, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.014379126951098442, + "learning_rate": 0.0005597194107104137, + "loss": 4.41373872756958, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.015897583216428757, + "learning_rate": 0.000559585254906272, + "loss": 4.377284049987793, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.016513120383024216, + "learning_rate": 0.000559450892376025, + "loss": 4.41200065612793, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.016625918447971344, + "learning_rate": 0.000559316323227926, + "loss": 4.381187438964844, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.016028188169002533, + "learning_rate": 0.000559181547570395, + "loss": 4.353909015655518, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01682914048433304, + "learning_rate": 0.0005590465655120183, + "loss": 4.382326126098633, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.017078889533877373, + "learning_rate": 0.0005589113771615486, + "loss": 4.334949493408203, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.018182938918471336, + "learning_rate": 0.0005587759826279046, + "loss": 4.348130226135254, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01940886303782463, + "learning_rate": 0.0005586403820201713, + "loss": 4.339404106140137, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.020644692704081535, + "learning_rate": 0.0005585045754475995, + "loss": 4.358404159545898, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02433430217206478, + "learning_rate": 0.0005583685630196064, + "loss": 4.3660502433776855, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.027824537828564644, + "learning_rate": 0.0005582323448457744, + "loss": 4.34050178527832, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.02973582223057747, + "learning_rate": 0.0005580959210358523, + "loss": 4.348322868347168, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.030396128073334694, + "learning_rate": 0.0005579592916997542, + "loss": 4.356224060058594, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.029760530218482018, + "learning_rate": 0.0005578224569475599, + "loss": 4.357470989227295, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.02855013869702816, + "learning_rate": 0.0005576854168895147, + "loss": 4.335353374481201, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03376635164022446, + "learning_rate": 0.0005575481716360292, + "loss": 4.331794738769531, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.034431129693984985, + "learning_rate": 0.0005574107212976796, + "loss": 4.336880207061768, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.03524135425686836, + "learning_rate": 0.000557273065985207, + "loss": 4.31537389755249, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.03152058273553848, + "learning_rate": 0.0005571352058095179, + "loss": 4.37662410736084, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02878640592098236, + "learning_rate": 0.0005569971408816838, + "loss": 4.292388916015625, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.029663654044270515, + "learning_rate": 0.0005568588713129409, + "loss": 4.344346523284912, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.027895759791135788, + "learning_rate": 0.0005567203972146906, + "loss": 4.365072727203369, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.026465769857168198, + "learning_rate": 0.000556581718698499, + "loss": 4.334959506988525, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.03191852569580078, + "learning_rate": 0.000556442835876097, + "loss": 4.361069679260254, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.03535444661974907, + "learning_rate": 0.0005563037488593799, + "loss": 4.349765777587891, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03451581299304962, + "learning_rate": 0.0005561644577604074, + "loss": 4.343142509460449, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029810845851898193, + "learning_rate": 0.000556024962691404, + "loss": 4.307485580444336, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.027523217722773552, + "learning_rate": 0.0005558852637647584, + "loss": 4.366406440734863, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03043208085000515, + "learning_rate": 0.0005557453610930235, + "loss": 4.356929779052734, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.03058922104537487, + "learning_rate": 0.0005556052547889161, + "loss": 4.348942279815674, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.032052166759967804, + "learning_rate": 0.0005554649449653177, + "loss": 4.3496809005737305, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.03170694038271904, + "learning_rate": 0.0005553244317352731, + "loss": 4.315959930419922, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 4.4757890701293945, + "eval_runtime": 57.7451, + "eval_samples_per_second": 42.289, + "eval_steps_per_second": 1.333, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.029025450348854065, + "learning_rate": 0.0005551837152119915, + "loss": 4.288647174835205, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.029403962194919586, + "learning_rate": 0.0005550427955088455, + "loss": 4.288555145263672, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.03265019878745079, + "learning_rate": 0.0005549016727393715, + "loss": 4.292625904083252, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.0387202650308609, + "learning_rate": 0.0005547603470172697, + "loss": 4.238926887512207, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.04554206505417824, + "learning_rate": 0.0005546188184564036, + "loss": 4.29966402053833, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.036231476813554764, + "learning_rate": 0.0005544770871708004, + "loss": 4.304398536682129, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.038703009486198425, + "learning_rate": 0.0005543351532746501, + "loss": 4.2945356369018555, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03224619850516319, + "learning_rate": 0.0005541930168823067, + "loss": 4.261351585388184, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.029722856357693672, + "learning_rate": 0.0005540506781082864, + "loss": 4.303143501281738, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.02764045260846615, + "learning_rate": 0.0005539081370672695, + "loss": 4.243679523468018, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.0312200877815485, + "learning_rate": 0.0005537653938740984, + "loss": 4.279904365539551, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.027349065989255905, + "learning_rate": 0.0005536224486437788, + "loss": 4.270228385925293, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.030385233461856842, + "learning_rate": 0.0005534793014914793, + "loss": 4.301590442657471, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.03292274847626686, + "learning_rate": 0.0005533359525325307, + "loss": 4.301828384399414, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.03265977278351784, + "learning_rate": 0.0005531924018824266, + "loss": 4.286200523376465, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.03177836537361145, + "learning_rate": 0.0005530486496568236, + "loss": 4.296319007873535, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.028879934921860695, + "learning_rate": 0.0005529046959715397, + "loss": 4.261706829071045, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.02855812944471836, + "learning_rate": 0.0005527605409425562, + "loss": 4.273903846740723, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03037901036441326, + "learning_rate": 0.0005526161846860161, + "loss": 4.248080253601074, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.029279105365276337, + "learning_rate": 0.0005524716273182245, + "loss": 4.281311988830566, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.02592580020427704, + "learning_rate": 0.0005523268689556489, + "loss": 4.204848289489746, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.021899603307247162, + "learning_rate": 0.0005521819097149183, + "loss": 4.232453346252441, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.019704442471265793, + "learning_rate": 0.0005520367497128238, + "loss": 4.253462314605713, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.019550519064068794, + "learning_rate": 0.0005518913890663182, + "loss": 4.307737350463867, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.020320113748311996, + "learning_rate": 0.0005517458278925161, + "loss": 4.257411956787109, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02118559181690216, + "learning_rate": 0.0005516000663086932, + "loss": 4.262089729309082, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.02050616219639778, + "learning_rate": 0.0005514541044322872, + "loss": 4.2490644454956055, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019793255254626274, + "learning_rate": 0.0005513079423808967, + "loss": 4.267400741577148, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.020760254934430122, + "learning_rate": 0.0005511615802722823, + "loss": 4.2532429695129395, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.019524799659848213, + "learning_rate": 0.0005510150182243647, + "loss": 4.267422199249268, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.020166637375950813, + "learning_rate": 0.0005508682563552269, + "loss": 4.214296817779541, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.01964394748210907, + "learning_rate": 0.0005507212947831119, + "loss": 4.252087116241455, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01690433733165264, + "learning_rate": 0.0005505741336264239, + "loss": 4.232336044311523, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.017682762816548347, + "learning_rate": 0.0005504267730037283, + "loss": 4.213962554931641, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01688133366405964, + "learning_rate": 0.0005502792130337508, + "loss": 4.199461460113525, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.019166061654686928, + "learning_rate": 0.0005501314538353776, + "loss": 4.228936672210693, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.024065518751740456, + "learning_rate": 0.000549983495527656, + "loss": 4.214227676391602, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02518288791179657, + "learning_rate": 0.000549835338229793, + "loss": 4.256412029266357, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.026093967258930206, + "learning_rate": 0.0005496869820611563, + "loss": 4.216564178466797, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.023775635287165642, + "learning_rate": 0.0005495384271412739, + "loss": 4.226430892944336, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.018810907378792763, + "learning_rate": 0.0005493896735898338, + "loss": 4.190013885498047, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.018974872305989265, + "learning_rate": 0.000549240721526684, + "loss": 4.246231555938721, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.02069833129644394, + "learning_rate": 0.0005490915710718325, + "loss": 4.198457717895508, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.02097306028008461, + "learning_rate": 0.0005489422223454471, + "loss": 4.212507247924805, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022165542468428612, + "learning_rate": 0.0005487926754678556, + "loss": 4.2043843269348145, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.019068919122219086, + "learning_rate": 0.0005486429305595448, + "loss": 4.197460651397705, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01701812446117401, + "learning_rate": 0.0005484929877411618, + "loss": 4.220056533813477, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01629726029932499, + "learning_rate": 0.0005483428471335127, + "loss": 4.214868545532227, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018661806359887123, + "learning_rate": 0.000548192508857563, + "loss": 4.227047920227051, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.02005891129374504, + "learning_rate": 0.0005480419730344377, + "loss": 4.197265625, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.02591128647327423, + "learning_rate": 0.0005478912397854208, + "loss": 4.211041450500488, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.032659754157066345, + "learning_rate": 0.0005477403092319552, + "loss": 4.230195045471191, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.04057808965444565, + "learning_rate": 0.0005475891814956429, + "loss": 4.223843097686768, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.04103454574942589, + "learning_rate": 0.0005474378566982449, + "loss": 4.197145462036133, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.03810729831457138, + "learning_rate": 0.0005472863349616809, + "loss": 4.237957000732422, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.028838301077485085, + "learning_rate": 0.0005471346164080292, + "loss": 4.206722259521484, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02436191774904728, + "learning_rate": 0.0005469827011595266, + "loss": 4.1983137130737305, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.022839387878775597, + "learning_rate": 0.0005468305893385686, + "loss": 4.20530366897583, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.02375260926783085, + "learning_rate": 0.0005466782810677088, + "loss": 4.213202476501465, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.02177741937339306, + "learning_rate": 0.0005465257764696593, + "loss": 4.200539588928223, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.02351468615233898, + "learning_rate": 0.0005463730756672903, + "loss": 4.198091506958008, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.023375794291496277, + "learning_rate": 0.0005462201787836301, + "loss": 4.193244934082031, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.02479805052280426, + "learning_rate": 0.0005460670859418651, + "loss": 4.2250871658325195, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.026894841343164444, + "learning_rate": 0.000545913797265339, + "loss": 4.241791248321533, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.030582403764128685, + "learning_rate": 0.0005457603128775541, + "loss": 4.203666687011719, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.032222259789705276, + "learning_rate": 0.0005456066329021699, + "loss": 4.209629535675049, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.03181060031056404, + "learning_rate": 0.0005454527574630037, + "loss": 4.200459957122803, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02687576413154602, + "learning_rate": 0.0005452986866840298, + "loss": 4.211819648742676, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.025988668203353882, + "learning_rate": 0.0005451444206893806, + "loss": 4.198040962219238, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.02937682718038559, + "learning_rate": 0.0005449899596033453, + "loss": 4.222935676574707, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.03043951839208603, + "learning_rate": 0.0005448353035503704, + "loss": 4.2031755447387695, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.030401621013879776, + "learning_rate": 0.0005446804526550596, + "loss": 4.198380470275879, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 4.368444919586182, + "eval_runtime": 58.3009, + "eval_samples_per_second": 41.886, + "eval_steps_per_second": 1.321, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.03461585193872452, + "learning_rate": 0.0005445254070421732, + "loss": 4.147369384765625, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.039084456861019135, + "learning_rate": 0.0005443701668366288, + "loss": 4.124129295349121, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.04021848738193512, + "learning_rate": 0.0005442147321635007, + "loss": 4.127362251281738, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.041000768542289734, + "learning_rate": 0.00054405910314802, + "loss": 4.1432061195373535, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.04008613899350166, + "learning_rate": 0.0005439032799155737, + "loss": 4.169931888580322, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.039394307881593704, + "learning_rate": 0.000543747262591706, + "loss": 4.155813217163086, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.04374868795275688, + "learning_rate": 0.0005435910513021172, + "loss": 4.159891128540039, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0443582646548748, + "learning_rate": 0.0005434346461726641, + "loss": 4.157362937927246, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.04601646214723587, + "learning_rate": 0.0005432780473293593, + "loss": 4.152097225189209, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.04327242821455002, + "learning_rate": 0.0005431212548983718, + "loss": 4.170074462890625, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.0413462370634079, + "learning_rate": 0.0005429642690060262, + "loss": 4.156710147857666, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.036855220794677734, + "learning_rate": 0.0005428070897788034, + "loss": 4.138053894042969, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.03781217709183693, + "learning_rate": 0.0005426497173433398, + "loss": 4.172306060791016, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.03886301815509796, + "learning_rate": 0.0005424921518264275, + "loss": 4.135544776916504, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.03730994835495949, + "learning_rate": 0.0005423343933550143, + "loss": 4.171654224395752, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.03271950036287308, + "learning_rate": 0.0005421764420562032, + "loss": 4.169906139373779, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.03292673081159592, + "learning_rate": 0.0005420182980572527, + "loss": 4.163660049438477, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.029366906732320786, + "learning_rate": 0.0005418599614855768, + "loss": 4.161055088043213, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.023805437609553337, + "learning_rate": 0.0005417014324687442, + "loss": 4.172390937805176, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.023178137838840485, + "learning_rate": 0.000541542711134479, + "loss": 4.146788597106934, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.02403450198471546, + "learning_rate": 0.0005413837976106599, + "loss": 4.124364852905273, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.025009717792272568, + "learning_rate": 0.0005412246920253208, + "loss": 4.12275505065918, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.023431289941072464, + "learning_rate": 0.0005410653945066503, + "loss": 4.134739398956299, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.023803815245628357, + "learning_rate": 0.0005409059051829914, + "loss": 4.130463123321533, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.021219994872808456, + "learning_rate": 0.0005407462241828417, + "loss": 4.124577522277832, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.01899666152894497, + "learning_rate": 0.0005405863516348535, + "loss": 4.167086124420166, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.01889493502676487, + "learning_rate": 0.000540426287667833, + "loss": 4.152744293212891, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.019546914845705032, + "learning_rate": 0.000540266032410741, + "loss": 4.122655868530273, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02085448056459427, + "learning_rate": 0.0005401055859926923, + "loss": 4.141949653625488, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.019959403201937675, + "learning_rate": 0.0005399449485429555, + "loss": 4.131907939910889, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.01885392889380455, + "learning_rate": 0.0005397841201909535, + "loss": 4.118953227996826, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.018216131255030632, + "learning_rate": 0.0005396231010662627, + "loss": 4.096459865570068, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.019385207444429398, + "learning_rate": 0.0005394618912986132, + "loss": 4.111050128936768, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.018676603212952614, + "learning_rate": 0.0005393004910178889, + "loss": 4.156215667724609, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.01876339502632618, + "learning_rate": 0.0005391389003541271, + "loss": 4.1107282638549805, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.01831623539328575, + "learning_rate": 0.0005389771194375185, + "loss": 4.154116630554199, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.018896661698818207, + "learning_rate": 0.0005388151483984071, + "loss": 4.093320846557617, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.019582601264119148, + "learning_rate": 0.0005386529873672898, + "loss": 4.106500625610352, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.020313601940870285, + "learning_rate": 0.0005384906364748172, + "loss": 4.071961879730225, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.020385954529047012, + "learning_rate": 0.000538328095851792, + "loss": 4.129006862640381, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02231123298406601, + "learning_rate": 0.0005381653656291707, + "loss": 4.067756652832031, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.022957179695367813, + "learning_rate": 0.0005380024459380619, + "loss": 4.088230133056641, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021438462659716606, + "learning_rate": 0.000537839336909727, + "loss": 4.128036022186279, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01893841102719307, + "learning_rate": 0.00053767603867558, + "loss": 4.155407905578613, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.018539676442742348, + "learning_rate": 0.0005375125513671874, + "loss": 4.101880073547363, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.01881185919046402, + "learning_rate": 0.0005373488751162681, + "loss": 4.118906021118164, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.021657396107912064, + "learning_rate": 0.0005371850100546929, + "loss": 4.091935157775879, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.023293187841773033, + "learning_rate": 0.000537020956314485, + "loss": 4.105803489685059, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.021306317299604416, + "learning_rate": 0.0005368567140278196, + "loss": 4.0610809326171875, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.022078553214669228, + "learning_rate": 0.0005366922833270235, + "loss": 4.09867525100708, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.02266664244234562, + "learning_rate": 0.0005365276643445757, + "loss": 4.1027326583862305, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.02091757394373417, + "learning_rate": 0.0005363628572131069, + "loss": 4.080266952514648, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.020374659448862076, + "learning_rate": 0.0005361978620653991, + "loss": 4.088479042053223, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.02294750325381756, + "learning_rate": 0.0005360326790343858, + "loss": 4.084364891052246, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.02337353676557541, + "learning_rate": 0.000535867308253152, + "loss": 4.125696182250977, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.022643696516752243, + "learning_rate": 0.0005357017498549341, + "loss": 4.097160339355469, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.021080605685710907, + "learning_rate": 0.0005355360039731196, + "loss": 4.102363109588623, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.02269427664577961, + "learning_rate": 0.0005353700707412466, + "loss": 4.089844703674316, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.02195413038134575, + "learning_rate": 0.0005352039502930048, + "loss": 4.132293701171875, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.01805616170167923, + "learning_rate": 0.0005350376427622343, + "loss": 4.0842814445495605, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.018048783764243126, + "learning_rate": 0.0005348711482829261, + "loss": 4.109591007232666, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.0199508648365736, + "learning_rate": 0.000534704466989222, + "loss": 4.0908074378967285, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.020186785608530045, + "learning_rate": 0.0005345375990154138, + "loss": 4.077959060668945, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.022526193410158157, + "learning_rate": 0.0005343705444959441, + "loss": 4.109979629516602, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.025224441662430763, + "learning_rate": 0.000534203303565406, + "loss": 4.047533988952637, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.027292724698781967, + "learning_rate": 0.000534035876358542, + "loss": 4.09566593170166, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.025669090449810028, + "learning_rate": 0.0005338682630102454, + "loss": 4.0606160163879395, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02791150100529194, + "learning_rate": 0.0005337004636555595, + "loss": 4.097081184387207, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.03294133022427559, + "learning_rate": 0.0005335324784296768, + "loss": 4.121442794799805, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.027322066947817802, + "learning_rate": 0.0005333643074679403, + "loss": 4.1188740730285645, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.024866372346878052, + "learning_rate": 0.000533195950905842, + "loss": 4.082042694091797, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.02901027724146843, + "learning_rate": 0.0005330274088790239, + "loss": 4.110199451446533, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 4.283970355987549, + "eval_runtime": 58.4003, + "eval_samples_per_second": 41.815, + "eval_steps_per_second": 1.318, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.028958842158317566, + "learning_rate": 0.0005328586815232773, + "loss": 4.040733337402344, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.03491180017590523, + "learning_rate": 0.0005326897689745428, + "loss": 4.0473432540893555, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.04025818780064583, + "learning_rate": 0.0005325206713689101, + "loss": 4.021976470947266, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03998899459838867, + "learning_rate": 0.0005323513888426181, + "loss": 4.04901123046875, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.042165111750364304, + "learning_rate": 0.0005321819215320546, + "loss": 4.050230979919434, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.03511597216129303, + "learning_rate": 0.0005320122695737565, + "loss": 4.0989990234375, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.04064979776740074, + "learning_rate": 0.0005318424331044094, + "loss": 4.051448822021484, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.047536592930555344, + "learning_rate": 0.0005316724122608473, + "loss": 4.064278602600098, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.05592669919133186, + "learning_rate": 0.0005315022071800527, + "loss": 4.045516014099121, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.04262633994221687, + "learning_rate": 0.0005313318179991574, + "loss": 4.060666084289551, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03997781500220299, + "learning_rate": 0.0005311612448554402, + "loss": 4.065315246582031, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.038152433931827545, + "learning_rate": 0.0005309904878863293, + "loss": 4.069721221923828, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.0367623008787632, + "learning_rate": 0.0005308195472294001, + "loss": 4.080986976623535, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.0358598455786705, + "learning_rate": 0.0005306484230223769, + "loss": 4.065290927886963, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.039188165217638016, + "learning_rate": 0.000530477115403131, + "loss": 4.024333477020264, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.036839909851551056, + "learning_rate": 0.0005303056245096817, + "loss": 4.067532539367676, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.03456985577940941, + "learning_rate": 0.0005301339504801967, + "loss": 4.048443794250488, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.039487943053245544, + "learning_rate": 0.0005299620934529902, + "loss": 4.0585737228393555, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.03415827825665474, + "learning_rate": 0.0005297900535665244, + "loss": 4.06104850769043, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.03209620714187622, + "learning_rate": 0.0005296178309594087, + "loss": 4.009165287017822, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.028947360813617706, + "learning_rate": 0.0005294454257704, + "loss": 4.067384243011475, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026836825534701347, + "learning_rate": 0.0005292728381384018, + "loss": 4.046419620513916, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024095216765999794, + "learning_rate": 0.0005291000682024647, + "loss": 4.019008636474609, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.020965851843357086, + "learning_rate": 0.0005289271161017867, + "loss": 4.020613670349121, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.02250855043530464, + "learning_rate": 0.0005287539819757119, + "loss": 4.031634330749512, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.021385207772254944, + "learning_rate": 0.0005285806659637315, + "loss": 4.086246013641357, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.021104589104652405, + "learning_rate": 0.000528407168205483, + "loss": 4.033970355987549, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.019584037363529205, + "learning_rate": 0.0005282334888407503, + "loss": 4.025118827819824, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.02107258513569832, + "learning_rate": 0.000528059628009464, + "loss": 4.020823955535889, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.020438389852643013, + "learning_rate": 0.0005278855858517003, + "loss": 4.016323566436768, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.01881488971412182, + "learning_rate": 0.0005277113625076822, + "loss": 3.9991841316223145, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.01763618178665638, + "learning_rate": 0.000527536958117778, + "loss": 4.0062150955200195, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.017308862879872322, + "learning_rate": 0.0005273623728225022, + "loss": 3.946723461151123, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01706845872104168, + "learning_rate": 0.0005271876067625153, + "loss": 4.056562900543213, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.018664294853806496, + "learning_rate": 0.0005270126600786228, + "loss": 4.024652481079102, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.018523404374718666, + "learning_rate": 0.0005268375329117762, + "loss": 3.962064743041992, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.017408935353159904, + "learning_rate": 0.0005266622254030724, + "loss": 4.008697509765625, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.01588570512831211, + "learning_rate": 0.0005264867376937534, + "loss": 4.011664390563965, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.015084310434758663, + "learning_rate": 0.0005263110699252065, + "loss": 3.996224880218506, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01637054979801178, + "learning_rate": 0.000526135222238964, + "loss": 4.015972137451172, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.01673802360892296, + "learning_rate": 0.0005259591947767032, + "loss": 3.975414276123047, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.016544019803404808, + "learning_rate": 0.0005257829876802464, + "loss": 4.012955188751221, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.017455460503697395, + "learning_rate": 0.0005256066010915603, + "loss": 4.010180473327637, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.017625460401177406, + "learning_rate": 0.0005254300351527566, + "loss": 4.001679420471191, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.01732795126736164, + "learning_rate": 0.000525253290006091, + "loss": 4.006107330322266, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.016654646024107933, + "learning_rate": 0.0005250763657939641, + "loss": 4.045180797576904, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.015085658989846706, + "learning_rate": 0.0005248992626589203, + "loss": 3.994246482849121, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.015112568624317646, + "learning_rate": 0.0005247219807436489, + "loss": 4.004963397979736, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.014882124960422516, + "learning_rate": 0.000524544520190982, + "loss": 4.029565811157227, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.016012471169233322, + "learning_rate": 0.000524366881143897, + "loss": 4.018208026885986, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.015761924907565117, + "learning_rate": 0.0005241890637455141, + "loss": 3.9911205768585205, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.017770355567336082, + "learning_rate": 0.0005240110681390978, + "loss": 4.028733253479004, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01959751546382904, + "learning_rate": 0.0005238328944680558, + "loss": 4.003322124481201, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.018713824450969696, + "learning_rate": 0.0005236545428759394, + "loss": 3.991718292236328, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.01992208883166313, + "learning_rate": 0.0005234760135064434, + "loss": 4.002507209777832, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.021671291440725327, + "learning_rate": 0.0005232973065034055, + "loss": 4.021344184875488, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024282798171043396, + "learning_rate": 0.0005231184220108069, + "loss": 3.9899158477783203, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.023294344544410706, + "learning_rate": 0.0005229393601727713, + "loss": 4.046627044677734, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.02029707282781601, + "learning_rate": 0.000522760121133566, + "loss": 4.022454261779785, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.02151239849627018, + "learning_rate": 0.0005225807050376004, + "loss": 4.005476951599121, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.02320881187915802, + "learning_rate": 0.0005224011120294269, + "loss": 4.012726783752441, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.026484191417694092, + "learning_rate": 0.0005222213422537401, + "loss": 3.9904391765594482, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.026279661804437637, + "learning_rate": 0.0005220413958553777, + "loss": 3.9738848209381104, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.024299729615449905, + "learning_rate": 0.0005218612729793188, + "loss": 4.003316402435303, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.02337595447897911, + "learning_rate": 0.0005216809737706856, + "loss": 3.954984188079834, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.02416428178548813, + "learning_rate": 0.0005215004983747415, + "loss": 4.0317063331604, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.022488107904791832, + "learning_rate": 0.0005213198469368925, + "loss": 3.9736647605895996, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.019700225442647934, + "learning_rate": 0.0005211390196026861, + "loss": 3.979867458343506, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.020382333546876907, + "learning_rate": 0.0005209580165178117, + "loss": 3.995251178741455, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.020149298012256622, + "learning_rate": 0.0005207768378281003, + "loss": 3.9987268447875977, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.019790509715676308, + "learning_rate": 0.0005205954836795241, + "loss": 3.9838991165161133, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.022096451371908188, + "learning_rate": 0.000520413954218197, + "loss": 3.995800256729126, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.214347839355469, + "eval_runtime": 58.8872, + "eval_samples_per_second": 41.469, + "eval_steps_per_second": 1.308, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02203402668237686, + "learning_rate": 0.0005202322495903739, + "loss": 3.915332078933716, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.02336675301194191, + "learning_rate": 0.0005200503699424509, + "loss": 3.914801597595215, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.027339637279510498, + "learning_rate": 0.0005198683154209654, + "loss": 3.9277780055999756, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.026872653514146805, + "learning_rate": 0.0005196860861725952, + "loss": 3.9160094261169434, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.026558255776762962, + "learning_rate": 0.0005195036823441592, + "loss": 3.9084300994873047, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.02644367143511772, + "learning_rate": 0.0005193211040826169, + "loss": 3.9328718185424805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.024677276611328125, + "learning_rate": 0.0005191383515350681, + "loss": 3.88993501663208, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.023194244131445885, + "learning_rate": 0.0005189554248487536, + "loss": 3.923567056655884, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.024653147906064987, + "learning_rate": 0.0005187723241710539, + "loss": 3.8740999698638916, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.026041202247142792, + "learning_rate": 0.0005185890496494899, + "loss": 3.9542148113250732, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.029586758464574814, + "learning_rate": 0.0005184056014317225, + "loss": 3.9241347312927246, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.0325787179172039, + "learning_rate": 0.0005182219796655528, + "loss": 3.9656593799591064, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.03183380514383316, + "learning_rate": 0.0005180381844989214, + "loss": 3.950648307800293, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.03434471786022186, + "learning_rate": 0.0005178542160799088, + "loss": 3.9590439796447754, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.030099432915449142, + "learning_rate": 0.0005176700745567353, + "loss": 3.9518113136291504, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.025950564071536064, + "learning_rate": 0.0005174857600777599, + "loss": 3.8908939361572266, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.029959935694932938, + "learning_rate": 0.0005173012727914816, + "loss": 3.945037364959717, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.029551763087511063, + "learning_rate": 0.0005171166128465389, + "loss": 3.923959732055664, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.03538523241877556, + "learning_rate": 0.0005169317803917083, + "loss": 3.9395527839660645, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.037924155592918396, + "learning_rate": 0.0005167467755759065, + "loss": 3.945115089416504, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.036773014813661575, + "learning_rate": 0.0005165615985481881, + "loss": 3.9484646320343018, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.04125838354229927, + "learning_rate": 0.0005163762494577474, + "loss": 4.000699043273926, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.044125135987997055, + "learning_rate": 0.0005161907284539165, + "loss": 3.9828028678894043, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.040853384882211685, + "learning_rate": 0.0005160050356861663, + "loss": 3.957505702972412, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.03501008450984955, + "learning_rate": 0.0005158191713041063, + "loss": 3.9360849857330322, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.04082556813955307, + "learning_rate": 0.0005156331354574838, + "loss": 3.95941424369812, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.045432962477207184, + "learning_rate": 0.0005154469282961847, + "loss": 3.9666922092437744, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.04048244655132294, + "learning_rate": 0.0005152605499702328, + "loss": 3.984555244445801, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.031016314402222633, + "learning_rate": 0.0005150740006297896, + "loss": 3.945204257965088, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.03210771828889847, + "learning_rate": 0.0005148872804251548, + "loss": 3.976186752319336, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.03130023553967476, + "learning_rate": 0.000514700389506765, + "loss": 3.915727138519287, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.031193193048238754, + "learning_rate": 0.0005145133280251954, + "loss": 3.961866855621338, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.027518954128026962, + "learning_rate": 0.0005143260961311578, + "loss": 3.9742956161499023, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.02751155197620392, + "learning_rate": 0.0005141386939755014, + "loss": 3.949662208557129, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.026033353060483932, + "learning_rate": 0.0005139511217092129, + "loss": 3.9327926635742188, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.02579621411859989, + "learning_rate": 0.000513763379483416, + "loss": 3.9334418773651123, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.023009559139609337, + "learning_rate": 0.0005135754674493709, + "loss": 3.969707489013672, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.02196447364985943, + "learning_rate": 0.0005133873857584753, + "loss": 3.9084434509277344, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.02206484042108059, + "learning_rate": 0.0005131991345622629, + "loss": 3.922117233276367, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.023360205814242363, + "learning_rate": 0.0005130107140124046, + "loss": 3.965158224105835, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02162669226527214, + "learning_rate": 0.0005128221242607072, + "loss": 3.9519729614257812, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.019719846546649933, + "learning_rate": 0.0005126333654591141, + "loss": 3.9195961952209473, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.018988260999321938, + "learning_rate": 0.0005124444377597049, + "loss": 3.937926769256592, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.019278401508927345, + "learning_rate": 0.0005122553413146951, + "loss": 3.9126808643341064, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019101005047559738, + "learning_rate": 0.0005120660762764366, + "loss": 3.8993749618530273, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.019156700000166893, + "learning_rate": 0.0005118766427974166, + "loss": 3.9531068801879883, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019944407045841217, + "learning_rate": 0.0005116870410302582, + "loss": 3.9645180702209473, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.01826174184679985, + "learning_rate": 0.0005114972711277202, + "loss": 3.948477268218994, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.0191368218511343, + "learning_rate": 0.0005113073332426968, + "loss": 3.944648027420044, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.017768343910574913, + "learning_rate": 0.0005111172275282174, + "loss": 3.9584474563598633, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.01955590397119522, + "learning_rate": 0.0005109269541374469, + "loss": 3.8936800956726074, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.02100536972284317, + "learning_rate": 0.000510736513223685, + "loss": 3.93919038772583, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.020873935893177986, + "learning_rate": 0.0005105459049403665, + "loss": 3.929244041442871, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.021876046434044838, + "learning_rate": 0.0005103551294410614, + "loss": 3.9264731407165527, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.021726684644818306, + "learning_rate": 0.0005101641868794736, + "loss": 3.923135280609131, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.019308188930153847, + "learning_rate": 0.0005099730774094423, + "loss": 3.9062135219573975, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01917017623782158, + "learning_rate": 0.0005097818011849408, + "loss": 3.898810386657715, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017933931201696396, + "learning_rate": 0.0005095903583600772, + "loss": 3.8899734020233154, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.015131795778870583, + "learning_rate": 0.0005093987490890933, + "loss": 3.9452900886535645, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.017190825194120407, + "learning_rate": 0.0005092069735263651, + "loss": 3.9124746322631836, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.01618487387895584, + "learning_rate": 0.000509015031826403, + "loss": 3.91654109954834, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.014920198358595371, + "learning_rate": 0.0005088229241438509, + "loss": 3.9144535064697266, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01615205965936184, + "learning_rate": 0.0005086306506334864, + "loss": 3.917003631591797, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.015370192006230354, + "learning_rate": 0.0005084382114502208, + "loss": 3.8910603523254395, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.015290996059775352, + "learning_rate": 0.0005082456067490988, + "loss": 3.9122695922851562, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.015913106501102448, + "learning_rate": 0.0005080528366852987, + "loss": 3.9304556846618652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.01639600843191147, + "learning_rate": 0.0005078599014141318, + "loss": 3.8974502086639404, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.017010601237416267, + "learning_rate": 0.0005076668010910425, + "loss": 3.9723148345947266, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.017778996378183365, + "learning_rate": 0.0005074735358716083, + "loss": 3.8905391693115234, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.020029611885547638, + "learning_rate": 0.0005072801059115394, + "loss": 3.934976100921631, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.01931857503950596, + "learning_rate": 0.000507086511366679, + "loss": 3.9152774810791016, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019750485196709633, + "learning_rate": 0.0005068927523930027, + "loss": 3.877650499343872, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.161914825439453, + "eval_runtime": 58.5217, + "eval_samples_per_second": 41.728, + "eval_steps_per_second": 1.316, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02073049545288086, + "learning_rate": 0.0005066988291466184, + "loss": 3.8331246376037598, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02272389829158783, + "learning_rate": 0.0005065047417837669, + "loss": 3.8847155570983887, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02139078453183174, + "learning_rate": 0.0005063104904608206, + "loss": 3.8276169300079346, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02043229155242443, + "learning_rate": 0.0005061160753342843, + "loss": 3.8804001808166504, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.02345564216375351, + "learning_rate": 0.0005059214965607948, + "loss": 3.851609468460083, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.023448185995221138, + "learning_rate": 0.0005057267542971207, + "loss": 3.8446507453918457, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.024826066568493843, + "learning_rate": 0.0005055318487001624, + "loss": 3.86712384223938, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.02491980977356434, + "learning_rate": 0.0005053367799269518, + "loss": 3.8374528884887695, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.023002834990620613, + "learning_rate": 0.0005051415481346522, + "loss": 3.783214569091797, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.02123952843248844, + "learning_rate": 0.0005049461534805585, + "loss": 3.8463428020477295, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.01981225423514843, + "learning_rate": 0.0005047505961220964, + "loss": 3.8307082653045654, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01915367692708969, + "learning_rate": 0.0005045548762168231, + "loss": 3.8190226554870605, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.01890859752893448, + "learning_rate": 0.0005043589939224266, + "loss": 3.8428220748901367, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.019150376319885254, + "learning_rate": 0.0005041629493967257, + "loss": 3.830247402191162, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.020099453628063202, + "learning_rate": 0.0005039667427976697, + "loss": 3.838806629180908, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.021136781200766563, + "learning_rate": 0.0005037703742833391, + "loss": 3.8344273567199707, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.022857150062918663, + "learning_rate": 0.0005035738440119442, + "loss": 3.8317880630493164, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.024891739711165428, + "learning_rate": 0.0005033771521418261, + "loss": 3.837855577468872, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.02595464698970318, + "learning_rate": 0.0005031802988314557, + "loss": 3.846050262451172, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.02722409926354885, + "learning_rate": 0.0005029832842394342, + "loss": 3.8164305686950684, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.024486860260367393, + "learning_rate": 0.0005027861085244927, + "loss": 3.8099875450134277, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.021551169455051422, + "learning_rate": 0.0005025887718454922, + "loss": 3.825925827026367, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.02121506817638874, + "learning_rate": 0.0005023912743614232, + "loss": 3.870727777481079, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.01948230341076851, + "learning_rate": 0.0005021936162314058, + "loss": 3.8035314083099365, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.01896044984459877, + "learning_rate": 0.0005019957976146898, + "loss": 3.84665584564209, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.020620400086045265, + "learning_rate": 0.0005017978186706538, + "loss": 3.8338310718536377, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.020855262875556946, + "learning_rate": 0.0005015996795588062, + "loss": 3.8781003952026367, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.02153870463371277, + "learning_rate": 0.0005014013804387838, + "loss": 3.8783490657806396, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.023751024156808853, + "learning_rate": 0.0005012029214703529, + "loss": 3.846198558807373, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.023987066000699997, + "learning_rate": 0.000501004302813408, + "loss": 3.834296226501465, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.023626182228326797, + "learning_rate": 0.0005008055246279727, + "loss": 3.8591227531433105, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.025707192718982697, + "learning_rate": 0.000500606587074199, + "loss": 3.8766489028930664, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.027723029255867004, + "learning_rate": 0.0005004074903123674, + "loss": 3.88264799118042, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.0278895553201437, + "learning_rate": 0.0005002082345028864, + "loss": 3.8932995796203613, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.025622662156820297, + "learning_rate": 0.0005000088198062929, + "loss": 3.8734612464904785, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.02639615908265114, + "learning_rate": 0.0004998092463832516, + "loss": 3.786813735961914, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.026765599846839905, + "learning_rate": 0.0004996095143945554, + "loss": 3.853966474533081, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02823319099843502, + "learning_rate": 0.0004994096240011246, + "loss": 3.8321340084075928, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.023394955322146416, + "learning_rate": 0.0004992095753640072, + "loss": 3.8527331352233887, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.023874662816524506, + "learning_rate": 0.0004990093686443791, + "loss": 3.8822879791259766, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.02301311306655407, + "learning_rate": 0.000498809004003543, + "loss": 3.846186876296997, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.022630592808127403, + "learning_rate": 0.0004986084816029291, + "loss": 3.867992639541626, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.02424517087638378, + "learning_rate": 0.0004984078016040946, + "loss": 3.867293357849121, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.024116624146699905, + "learning_rate": 0.000498206964168724, + "loss": 3.846282958984375, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.02187768742442131, + "learning_rate": 0.000498005969458628, + "loss": 3.8429484367370605, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.01879809983074665, + "learning_rate": 0.0004978048176357447, + "loss": 3.863215446472168, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.020426234230399132, + "learning_rate": 0.0004976035088621385, + "loss": 3.825230121612549, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01992812193930149, + "learning_rate": 0.0004974020433000003, + "loss": 3.851665496826172, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.020693328231573105, + "learning_rate": 0.000497200421111647, + "loss": 3.906999111175537, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.021137919276952744, + "learning_rate": 0.0004969986424595221, + "loss": 3.8369035720825195, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.023285407572984695, + "learning_rate": 0.0004967967075061949, + "loss": 3.8658390045166016, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.0219135619699955, + "learning_rate": 0.0004965946164143609, + "loss": 3.873072862625122, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.020101983100175858, + "learning_rate": 0.0004963923693468413, + "loss": 3.847805976867676, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.01875176839530468, + "learning_rate": 0.0004961899664665827, + "loss": 3.7976622581481934, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01913360133767128, + "learning_rate": 0.0004959874079366576, + "loss": 3.853851318359375, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.016863619908690453, + "learning_rate": 0.0004957846939202637, + "loss": 3.877378225326538, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.017802366986870766, + "learning_rate": 0.000495581824580724, + "loss": 3.861356019973755, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.01981814205646515, + "learning_rate": 0.0004953788000814866, + "loss": 3.88613224029541, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.017508897930383682, + "learning_rate": 0.000495175620586125, + "loss": 3.8202030658721924, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.019240548834204674, + "learning_rate": 0.0004949722862583368, + "loss": 3.853827714920044, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.022305598482489586, + "learning_rate": 0.000494768797261945, + "loss": 3.885148763656616, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.02531025931239128, + "learning_rate": 0.000494565153760897, + "loss": 3.8412041664123535, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.024474849924445152, + "learning_rate": 0.0004943613559192648, + "loss": 3.865644931793213, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.020430924370884895, + "learning_rate": 0.0004941574039012444, + "loss": 3.8649442195892334, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.01906431093811989, + "learning_rate": 0.0004939532978711566, + "loss": 3.8462204933166504, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.021118057891726494, + "learning_rate": 0.0004937490379934456, + "loss": 3.8652522563934326, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.020337693393230438, + "learning_rate": 0.00049354462443268, + "loss": 3.8506221771240234, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.019090797752141953, + "learning_rate": 0.0004933400573535521, + "loss": 3.881237506866455, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.020973797887563705, + "learning_rate": 0.000493135336920878, + "loss": 3.821274757385254, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.021984362974762917, + "learning_rate": 0.0004929304632995973, + "loss": 3.8793835639953613, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.0215684212744236, + "learning_rate": 0.0004927254366547727, + "loss": 3.8259663581848145, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.021673865616321564, + "learning_rate": 0.0004925202571515908, + "loss": 3.8505711555480957, + "step": 1080 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 4.604420041998336e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-1080/training_args.bin b/runs/l2r90-baseline/checkpoint-1080/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1080/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-1440/chat_template.jinja b/runs/l2r90-baseline/checkpoint-1440/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-1440/config.json b/runs/l2r90-baseline/checkpoint-1440/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-1440/generation_config.json b/runs/l2r90-baseline/checkpoint-1440/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-1440/model.safetensors b/runs/l2r90-baseline/checkpoint-1440/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ad8bc1822a865e8b9ac78873d3bab8e02d6f7e34 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf794842ba70c1de9feac51edce37557dbe9c698c7740fbd60882205103446e9 +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-1440/optimizer.pt b/runs/l2r90-baseline/checkpoint-1440/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..0c42825e08a66be2721996eb54e1f752527eb96f --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d1ebc2f7d7a4cd34e2ea6acaa33dad13d10e049640e28cebc20e90657eca9b2 +size 1166825338 diff --git a/runs/l2r90-baseline/checkpoint-1440/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-1440/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..aeac82010ce2bacd37bd6af051976d38c4fe2812 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b1dfed66aea4e59c5fbb97f6f7b9eedaa2102bc813da51af9a94ba9c42b0e8c +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1440/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-1440/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..f1f618778c4f7500b21ac62d13c3ef2bebccd4b9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31bbc97cda656d11c5036ea59105511b2f07d7b2834b974ae356d2de6f8d922d +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1440/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-1440/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..5fdc074e62a1d8319b7a92b775a1095e5ea19aad --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f54eefd68dc84d612a90080eb0724dd4cdddd67659a3ca9aa7aeb336ba20c1fe +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1440/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-1440/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..ffe532f21731696b54618fdc406075cce9e04300 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37d7701ca99aff6a37aa49666cef9609252548ec1ef723213fedc3c95ad648af +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1440/scheduler.pt b/runs/l2r90-baseline/checkpoint-1440/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..59b8db4a485766e1d20d11a905a37774da0cec09 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fe49e98b5d706cd9174dc7cf5e5cd5039ca03537426b9020792a5363720022f +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-1440/tokenizer.json b/runs/l2r90-baseline/checkpoint-1440/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-1440/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-1440/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-1440/trainer_state.json b/runs/l2r90-baseline/checkpoint-1440/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..da1bc173cd768af9d922e1d8814377a8e81a2a22 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/trainer_state.json @@ -0,0 +1,10266 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 20.0, + "eval_steps": 500, + "global_step": 1440, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 5.549652099609375, + "eval_runtime": 57.761, + "eval_samples_per_second": 42.278, + "eval_steps_per_second": 1.333, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.03958690166473389, + "learning_rate": 0.0005919487226162231, + "loss": 5.459286689758301, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.03299890458583832, + "learning_rate": 0.0005918869546699877, + "loss": 5.45464563369751, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.028040863573551178, + "learning_rate": 0.0005918249539727509, + "loss": 5.444796562194824, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.028365720063447952, + "learning_rate": 0.0005917627205744655, + "loss": 5.447852611541748, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.02822224237024784, + "learning_rate": 0.0005917002545252719, + "loss": 5.453883171081543, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.036075398325920105, + "learning_rate": 0.0005916375558754977, + "loss": 5.462671279907227, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04683876410126686, + "learning_rate": 0.0005915746246756581, + "loss": 5.443917274475098, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.05457588657736778, + "learning_rate": 0.0005915114609764558, + "loss": 5.444374084472656, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.04250836744904518, + "learning_rate": 0.0005914480648287804, + "loss": 5.435029983520508, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.029477844014763832, + "learning_rate": 0.0005913844362837093, + "loss": 5.421862602233887, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.03673284500837326, + "learning_rate": 0.0005913205753925066, + "loss": 5.430391311645508, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.030489861965179443, + "learning_rate": 0.0005912564822066241, + "loss": 5.389314651489258, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.04037046805024147, + "learning_rate": 0.0005911921567777005, + "loss": 5.425060272216797, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.0445217490196228, + "learning_rate": 0.0005911275991575613, + "loss": 5.437565326690674, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.03885014355182648, + "learning_rate": 0.0005910628093982198, + "loss": 5.410930633544922, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.0334964245557785, + "learning_rate": 0.0005909977875518759, + "loss": 5.412092208862305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.03499286621809006, + "learning_rate": 0.0005909325336709164, + "loss": 5.382216453552246, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.04171644523739815, + "learning_rate": 0.0005908670478079152, + "loss": 5.387385368347168, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.057558078318834305, + "learning_rate": 0.000590801330015633, + "loss": 5.437925815582275, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.053846780210733414, + "learning_rate": 0.0005907353803470177, + "loss": 5.41644811630249, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.027259036898612976, + "learning_rate": 0.0005906691988552034, + "loss": 5.409185409545898, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.03394043818116188, + "learning_rate": 0.0005906027855935115, + "loss": 5.377973556518555, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.03731574863195419, + "learning_rate": 0.0005905361406154501, + "loss": 5.382505893707275, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.04705141484737396, + "learning_rate": 0.0005904692639747137, + "loss": 5.413661003112793, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.044811323285102844, + "learning_rate": 0.0005904021557251837, + "loss": 5.417835235595703, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.04000752419233322, + "learning_rate": 0.0005903348159209277, + "loss": 5.406160354614258, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.038364164531230927, + "learning_rate": 0.0005902672446162007, + "loss": 5.357950687408447, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.030500127002596855, + "learning_rate": 0.0005901994418654432, + "loss": 5.344951629638672, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.030644729733467102, + "learning_rate": 0.0005901314077232829, + "loss": 5.3599348068237305, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.037976790219545364, + "learning_rate": 0.0005900631422445335, + "loss": 5.354084491729736, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.04106011986732483, + "learning_rate": 0.0005899946454841955, + "loss": 5.363678455352783, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.04334832355380058, + "learning_rate": 0.0005899259174974553, + "loss": 5.350959300994873, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.049939028918743134, + "learning_rate": 0.000589856958339686, + "loss": 5.3317155838012695, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.05300058051943779, + "learning_rate": 0.0005897877680664465, + "loss": 5.353314399719238, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.05943184345960617, + "learning_rate": 0.0005897183467334822, + "loss": 5.3683929443359375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.05600789934396744, + "learning_rate": 0.0005896486943967249, + "loss": 5.345516204833984, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.05392718315124512, + "learning_rate": 0.0005895788111122919, + "loss": 5.291389465332031, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.037913475185632706, + "learning_rate": 0.0005895086969364871, + "loss": 5.353423118591309, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.03362123295664787, + "learning_rate": 0.0005894383519258001, + "loss": 5.379500389099121, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032558735460042953, + "learning_rate": 0.0005893677761369066, + "loss": 5.34043025970459, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.04074402526021004, + "learning_rate": 0.0005892969696266683, + "loss": 5.303869247436523, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.04270484298467636, + "learning_rate": 0.0005892259324521328, + "loss": 5.27297830581665, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.04604959115386009, + "learning_rate": 0.0005891546646705334, + "loss": 5.313268184661865, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.04462246224284172, + "learning_rate": 0.0005890831663392891, + "loss": 5.31167459487915, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.04966094344854355, + "learning_rate": 0.0005890114375160052, + "loss": 5.316695690155029, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.04662073776125908, + "learning_rate": 0.0005889394782584718, + "loss": 5.316898822784424, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.04380315542221069, + "learning_rate": 0.0005888672886246656, + "loss": 5.285087585449219, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03239579498767853, + "learning_rate": 0.0005887948686727483, + "loss": 5.304283142089844, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.029690183699131012, + "learning_rate": 0.0005887222184610675, + "loss": 5.305715084075928, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.03444967418909073, + "learning_rate": 0.0005886493380481559, + "loss": 5.2707672119140625, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.03302565962076187, + "learning_rate": 0.0005885762274927322, + "loss": 5.277251243591309, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.028477763757109642, + "learning_rate": 0.0005885028868537, + "loss": 5.295318603515625, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.022988801822066307, + "learning_rate": 0.0005884293161901487, + "loss": 5.249057292938232, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.026115216314792633, + "learning_rate": 0.000588355515561353, + "loss": 5.270636081695557, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.02576332725584507, + "learning_rate": 0.0005882814850267723, + "loss": 5.263096332550049, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.023978352546691895, + "learning_rate": 0.0005882072246460521, + "loss": 5.276605606079102, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.020695513114333153, + "learning_rate": 0.0005881327344790223, + "loss": 5.2141432762146, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01952219195663929, + "learning_rate": 0.0005880580145856986, + "loss": 5.214289665222168, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.022633206099271774, + "learning_rate": 0.0005879830650262813, + "loss": 5.236690998077393, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.023082347586750984, + "learning_rate": 0.0005879078858611557, + "loss": 5.2092156410217285, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.019620463252067566, + "learning_rate": 0.0005878324771508927, + "loss": 5.264451503753662, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.021504493430256844, + "learning_rate": 0.0005877568389562475, + "loss": 5.23444938659668, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.025672364979982376, + "learning_rate": 0.0005876809713381606, + "loss": 5.205289363861084, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.034436631947755814, + "learning_rate": 0.0005876048743577569, + "loss": 5.2047648429870605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.042389530688524246, + "learning_rate": 0.0005875285480763466, + "loss": 5.192008018493652, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.03310716897249222, + "learning_rate": 0.0005874519925554244, + "loss": 5.162734031677246, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.03029288910329342, + "learning_rate": 0.0005873752078566694, + "loss": 5.212151527404785, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04619310051202774, + "learning_rate": 0.0005872981940419459, + "loss": 5.215909004211426, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.05618087574839592, + "learning_rate": 0.0005872209511733027, + "loss": 5.210192680358887, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.03783707693219185, + "learning_rate": 0.0005871434793129726, + "loss": 5.201597690582275, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.03721104562282562, + "learning_rate": 0.0005870657785233735, + "loss": 5.188891410827637, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.033336155116558075, + "learning_rate": 0.0005869878488671075, + "loss": 5.1598615646362305, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.2592363357543945, + "eval_runtime": 57.3343, + "eval_samples_per_second": 42.592, + "eval_steps_per_second": 1.343, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03719741478562355, + "learning_rate": 0.0005869096904069611, + "loss": 5.1646623611450195, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.04888090118765831, + "learning_rate": 0.0005868313032059052, + "loss": 5.171182155609131, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.05073149502277374, + "learning_rate": 0.0005867526873270949, + "loss": 5.183949947357178, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.04187731072306633, + "learning_rate": 0.0005866738428338695, + "loss": 5.176133155822754, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.039445552974939346, + "learning_rate": 0.000586594769789753, + "loss": 5.148100852966309, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.035551708191633224, + "learning_rate": 0.0005865154682584524, + "loss": 5.151512145996094, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.031175289303064346, + "learning_rate": 0.0005864359383038602, + "loss": 5.102468967437744, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.03696789592504501, + "learning_rate": 0.000586356179990052, + "loss": 5.119902610778809, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.04836135357618332, + "learning_rate": 0.0005862761933812875, + "loss": 5.15183162689209, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.05206914246082306, + "learning_rate": 0.0005861959785420106, + "loss": 5.154870986938477, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.046368084847927094, + "learning_rate": 0.000586115535536849, + "loss": 5.130651473999023, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.07146432995796204, + "learning_rate": 0.000586034864430614, + "loss": 5.175804138183594, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.08809789270162582, + "learning_rate": 0.0005859539652883008, + "loss": 5.229716777801514, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.08116360008716583, + "learning_rate": 0.0005858728381750888, + "loss": 5.21910285949707, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.08282570540904999, + "learning_rate": 0.0005857914831563402, + "loss": 5.2336506843566895, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.07403547316789627, + "learning_rate": 0.0005857099002976016, + "loss": 5.251604080200195, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.07024004310369492, + "learning_rate": 0.0005856280896646026, + "loss": 5.267114162445068, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.056330062448978424, + "learning_rate": 0.0005855460513232567, + "loss": 5.201976299285889, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.051763955503702164, + "learning_rate": 0.0005854637853396606, + "loss": 5.1733198165893555, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.039900992065668106, + "learning_rate": 0.0005853812917800945, + "loss": 5.222941875457764, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.04187863692641258, + "learning_rate": 0.0005852985707110221, + "loss": 5.165033340454102, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.042645812034606934, + "learning_rate": 0.0005852156221990901, + "loss": 5.157554626464844, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.03856075555086136, + "learning_rate": 0.0005851324463111289, + "loss": 5.184629440307617, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.029387330636382103, + "learning_rate": 0.0005850490431141516, + "loss": 5.118739128112793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.027584781870245934, + "learning_rate": 0.0005849654126753545, + "loss": 5.100855350494385, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.023655924946069717, + "learning_rate": 0.0005848815550621175, + "loss": 5.154520034790039, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.024163050577044487, + "learning_rate": 0.0005847974703420028, + "loss": 5.105189323425293, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.020369920879602432, + "learning_rate": 0.0005847131585827564, + "loss": 5.113671779632568, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.021030737087130547, + "learning_rate": 0.0005846286198523061, + "loss": 5.0847649574279785, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.01729249767959118, + "learning_rate": 0.0005845438542187638, + "loss": 5.053225517272949, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.018704712390899658, + "learning_rate": 0.0005844588617504236, + "loss": 5.079106330871582, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.02049064077436924, + "learning_rate": 0.0005843736425157621, + "loss": 5.044185161590576, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.020736856386065483, + "learning_rate": 0.000584288196583439, + "loss": 5.041932106018066, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.016543276607990265, + "learning_rate": 0.0005842025240222966, + "loss": 5.032522201538086, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.018414106220006943, + "learning_rate": 0.0005841166249013598, + "loss": 5.037397384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.01611698977649212, + "learning_rate": 0.0005840304992898359, + "loss": 5.038179397583008, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.014310561120510101, + "learning_rate": 0.0005839441472571147, + "loss": 5.0146484375, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.015472214668989182, + "learning_rate": 0.0005838575688727685, + "loss": 5.060323715209961, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.016373401507735252, + "learning_rate": 0.0005837707642065518, + "loss": 5.032100677490234, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.018694665282964706, + "learning_rate": 0.000583683733328402, + "loss": 5.036993026733398, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.019977087154984474, + "learning_rate": 0.0005835964763084378, + "loss": 4.991037368774414, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.023741086944937706, + "learning_rate": 0.000583508993216961, + "loss": 5.043832778930664, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.025061562657356262, + "learning_rate": 0.0005834212841244548, + "loss": 5.016516208648682, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.020848384127020836, + "learning_rate": 0.0005833333491015851, + "loss": 4.986289978027344, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.015086354687809944, + "learning_rate": 0.0005832451882191995, + "loss": 4.9948625564575195, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.015096964314579964, + "learning_rate": 0.0005831568015483274, + "loss": 4.996342658996582, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016932200640439987, + "learning_rate": 0.0005830681891601807, + "loss": 4.998758316040039, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.019936930388212204, + "learning_rate": 0.0005829793511261525, + "loss": 4.959854602813721, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.027658682316541672, + "learning_rate": 0.0005828902875178183, + "loss": 4.968629837036133, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.03299275413155556, + "learning_rate": 0.0005828009984069347, + "loss": 4.994814872741699, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.033717650920152664, + "learning_rate": 0.0005827114838654406, + "loss": 4.976295471191406, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.02969924360513687, + "learning_rate": 0.0005826217439654562, + "loss": 4.969155788421631, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.02846827358007431, + "learning_rate": 0.0005825317787792831, + "loss": 4.989721298217773, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.03670600801706314, + "learning_rate": 0.0005824415883794049, + "loss": 4.988689422607422, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.045570675283670425, + "learning_rate": 0.0005823511728384863, + "loss": 4.974407196044922, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.04564111679792404, + "learning_rate": 0.0005822605322293733, + "loss": 4.966331481933594, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.04173210263252258, + "learning_rate": 0.0005821696666250937, + "loss": 4.925197601318359, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.04228242486715317, + "learning_rate": 0.0005820785760988559, + "loss": 4.965020656585693, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.05144782364368439, + "learning_rate": 0.0005819872607240503, + "loss": 5.0022149085998535, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.049400657415390015, + "learning_rate": 0.0005818957205742478, + "loss": 4.9635114669799805, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.043692708015441895, + "learning_rate": 0.0005818039557232005, + "loss": 4.977723121643066, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.04891742020845413, + "learning_rate": 0.0005817119662448421, + "loss": 4.996312141418457, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.035564981400966644, + "learning_rate": 0.0005816197522132866, + "loss": 4.940932273864746, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04679827764630318, + "learning_rate": 0.0005815273137028292, + "loss": 4.983131408691406, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.05489890277385712, + "learning_rate": 0.0005814346507879459, + "loss": 4.977827548980713, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04640809819102287, + "learning_rate": 0.0005813417635432937, + "loss": 4.954761505126953, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04439452290534973, + "learning_rate": 0.00058124865204371, + "loss": 4.914865970611572, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.04587096720933914, + "learning_rate": 0.0005811553163642131, + "loss": 4.949592590332031, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.044119734317064285, + "learning_rate": 0.0005810617565800019, + "loss": 4.956068992614746, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.043133001774549484, + "learning_rate": 0.0005809679727664559, + "loss": 4.993522644042969, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04315278306603432, + "learning_rate": 0.0005808739649991348, + "loss": 4.894291877746582, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04819472134113312, + "learning_rate": 0.0005807797333537792, + "loss": 4.958306789398193, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.023706436157227, + "eval_runtime": 57.4082, + "eval_samples_per_second": 42.537, + "eval_steps_per_second": 1.341, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.04403403028845787, + "learning_rate": 0.0005806852779063098, + "loss": 4.884971618652344, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.041993748396635056, + "learning_rate": 0.0005805905987328275, + "loss": 4.916353225708008, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.042302004992961884, + "learning_rate": 0.0005804956959096138, + "loss": 4.9012556076049805, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.044500596821308136, + "learning_rate": 0.0005804005695131299, + "loss": 4.915240287780762, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.03914736211299896, + "learning_rate": 0.0005803052196200177, + "loss": 4.885915756225586, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.0361703559756279, + "learning_rate": 0.0005802096463070988, + "loss": 4.829287528991699, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03432302549481392, + "learning_rate": 0.0005801138496513749, + "loss": 4.926834583282471, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.0317804180085659, + "learning_rate": 0.0005800178297300276, + "loss": 4.8707170486450195, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.035261351615190506, + "learning_rate": 0.0005799215866204183, + "loss": 4.897791862487793, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.03276067599654198, + "learning_rate": 0.0005798251204000886, + "loss": 4.864504814147949, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.034757066518068314, + "learning_rate": 0.0005797284311467594, + "loss": 4.890812873840332, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.033726081252098083, + "learning_rate": 0.0005796315189383316, + "loss": 4.831684112548828, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.03074726276099682, + "learning_rate": 0.0005795343838528855, + "loss": 4.8534111976623535, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02215893566608429, + "learning_rate": 0.0005794370259686811, + "loss": 4.8311238288879395, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.02132033370435238, + "learning_rate": 0.0005793394453641581, + "loss": 4.842617034912109, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020825816318392754, + "learning_rate": 0.0005792416421179351, + "loss": 4.84535026550293, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.019334357231855392, + "learning_rate": 0.0005791436163088106, + "loss": 4.839937686920166, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017372334375977516, + "learning_rate": 0.0005790453680157622, + "loss": 4.8115553855896, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.0190406683832407, + "learning_rate": 0.0005789468973179468, + "loss": 4.832542419433594, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.017372574657201767, + "learning_rate": 0.0005788482042947004, + "loss": 4.793632507324219, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.019171632826328278, + "learning_rate": 0.0005787492890255382, + "loss": 4.776987552642822, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.02109278179705143, + "learning_rate": 0.0005786501515901545, + "loss": 4.8216657638549805, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.026234494522213936, + "learning_rate": 0.0005785507920684225, + "loss": 4.8109049797058105, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.0357641875743866, + "learning_rate": 0.0005784512105403944, + "loss": 4.821109771728516, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.03627947345376015, + "learning_rate": 0.000578351407086301, + "loss": 4.810632228851318, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.032253146171569824, + "learning_rate": 0.0005782513817865527, + "loss": 4.7930521965026855, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.03402210772037506, + "learning_rate": 0.0005781511347217375, + "loss": 4.772500991821289, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.0321584977209568, + "learning_rate": 0.000578050665972623, + "loss": 4.781399726867676, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.030242707580327988, + "learning_rate": 0.0005779499756201549, + "loss": 4.802542686462402, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.02361336164176464, + "learning_rate": 0.0005778490637454576, + "loss": 4.771912574768066, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.028781302273273468, + "learning_rate": 0.000577747930429834, + "loss": 4.796924591064453, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.027042683213949203, + "learning_rate": 0.0005776465757547654, + "loss": 4.745218276977539, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.02715984545648098, + "learning_rate": 0.0005775449998019113, + "loss": 4.792944431304932, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.0314096100628376, + "learning_rate": 0.00057744320265311, + "loss": 4.7649993896484375, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.028749193996191025, + "learning_rate": 0.0005773411843903773, + "loss": 4.7980546951293945, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.02655309997498989, + "learning_rate": 0.0005772389450959075, + "loss": 4.746520519256592, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.027497397735714912, + "learning_rate": 0.000577136484852073, + "loss": 4.7490715980529785, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.0343393050134182, + "learning_rate": 0.000577033803741424, + "loss": 4.74436092376709, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.04444412514567375, + "learning_rate": 0.0005769309018466891, + "loss": 4.755973815917969, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.0437556728720665, + "learning_rate": 0.0005768277792507744, + "loss": 4.784368515014648, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0369921550154686, + "learning_rate": 0.0005767244360367638, + "loss": 4.753410339355469, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.05038141459226608, + "learning_rate": 0.0005766208722879192, + "loss": 4.755491256713867, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.04511422663927078, + "learning_rate": 0.00057651708808768, + "loss": 4.760293006896973, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.048762883991003036, + "learning_rate": 0.0005764130835196632, + "loss": 4.755958080291748, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.06749686598777771, + "learning_rate": 0.0005763088586676634, + "loss": 4.802557468414307, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.06451250612735748, + "learning_rate": 0.0005762044136156528, + "loss": 4.812925338745117, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.05436156690120697, + "learning_rate": 0.0005760997484477809, + "loss": 4.807103157043457, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.05233754590153694, + "learning_rate": 0.0005759948632483742, + "loss": 4.782449722290039, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.05519746616482735, + "learning_rate": 0.0005758897581019371, + "loss": 4.808449745178223, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.0578567236661911, + "learning_rate": 0.000575784433093151, + "loss": 4.7676005363464355, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.0588264986872673, + "learning_rate": 0.0005756788883068743, + "loss": 4.805377960205078, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04225790873169899, + "learning_rate": 0.0005755731238281423, + "loss": 4.748775005340576, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.0376368910074234, + "learning_rate": 0.0005754671397421678, + "loss": 4.759791374206543, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03830771893262863, + "learning_rate": 0.0005753609361343402, + "loss": 4.729974746704102, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.03684520348906517, + "learning_rate": 0.0005752545130902256, + "loss": 4.729854583740234, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02983408235013485, + "learning_rate": 0.0005751478706955674, + "loss": 4.78502082824707, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02854904904961586, + "learning_rate": 0.0005750410090362854, + "loss": 4.717472076416016, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.023819150403141975, + "learning_rate": 0.0005749339281984761, + "loss": 4.717124938964844, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.02477158047258854, + "learning_rate": 0.0005748266282684124, + "loss": 4.71781063079834, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.022002534940838814, + "learning_rate": 0.0005747191093325443, + "loss": 4.7480669021606445, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.02133483812212944, + "learning_rate": 0.0005746113714774976, + "loss": 4.693770408630371, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.022407138720154762, + "learning_rate": 0.0005745034147900747, + "loss": 4.720734119415283, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.020850926637649536, + "learning_rate": 0.0005743952393572544, + "loss": 4.745505332946777, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.020420582965016365, + "learning_rate": 0.0005742868452661918, + "loss": 4.688610076904297, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.019014107063412666, + "learning_rate": 0.0005741782326042181, + "loss": 4.673436164855957, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.017393039539456367, + "learning_rate": 0.0005740694014588403, + "loss": 4.6941118240356445, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.017877329140901566, + "learning_rate": 0.0005739603519177419, + "loss": 4.667474269866943, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.017749635502696037, + "learning_rate": 0.0005738510840687821, + "loss": 4.6901092529296875, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.018746376037597656, + "learning_rate": 0.000573741597999996, + "loss": 4.680166721343994, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016260217875242233, + "learning_rate": 0.0005736318937995947, + "loss": 4.6799821853637695, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015099814161658287, + "learning_rate": 0.0005735219715559646, + "loss": 4.680666923522949, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.016539735719561577, + "learning_rate": 0.0005734118313576683, + "loss": 4.639509201049805, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 4.768754482269287, + "eval_runtime": 58.3352, + "eval_samples_per_second": 41.862, + "eval_steps_per_second": 1.32, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.019733518362045288, + "learning_rate": 0.0005733014732934436, + "loss": 4.635312080383301, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.022571036592125893, + "learning_rate": 0.0005731908974522042, + "loss": 4.6041765213012695, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.022238552570343018, + "learning_rate": 0.0005730801039230389, + "loss": 4.648976802825928, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.02115180902183056, + "learning_rate": 0.0005729690927952123, + "loss": 4.623968124389648, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.021789928898215294, + "learning_rate": 0.0005728578641581637, + "loss": 4.620028018951416, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.02593887411057949, + "learning_rate": 0.0005727464181015081, + "loss": 4.591062545776367, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.0347842313349247, + "learning_rate": 0.0005726347547150357, + "loss": 4.632149696350098, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04119519516825676, + "learning_rate": 0.0005725228740887117, + "loss": 4.6254191398620605, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.04895750433206558, + "learning_rate": 0.0005724107763126761, + "loss": 4.621372222900391, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.05752161890268326, + "learning_rate": 0.0005722984614772442, + "loss": 4.61965274810791, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05115870013833046, + "learning_rate": 0.000572185929672906, + "loss": 4.63119649887085, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.05721089243888855, + "learning_rate": 0.0005720731809903263, + "loss": 4.652647018432617, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.0646839588880539, + "learning_rate": 0.000571960215520345, + "loss": 4.642184257507324, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.0682854950428009, + "learning_rate": 0.0005718470333539757, + "loss": 4.676589012145996, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.057402849197387695, + "learning_rate": 0.0005717336345824077, + "loss": 4.6585283279418945, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.05885373800992966, + "learning_rate": 0.0005716200192970043, + "loss": 4.671728134155273, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.05265705659985542, + "learning_rate": 0.0005715061875893032, + "loss": 4.641249656677246, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.05417842045426369, + "learning_rate": 0.0005713921395510166, + "loss": 4.661381721496582, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.06602369248867035, + "learning_rate": 0.0005712778752740307, + "loss": 4.684137344360352, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.06004182621836662, + "learning_rate": 0.0005711633948504066, + "loss": 4.625332832336426, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.04801511764526367, + "learning_rate": 0.0005710486983723787, + "loss": 4.664791107177734, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.04380105435848236, + "learning_rate": 0.0005709337859323561, + "loss": 4.664966583251953, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.03858727216720581, + "learning_rate": 0.0005708186576229218, + "loss": 4.661867618560791, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03240576758980751, + "learning_rate": 0.0005707033135368323, + "loss": 4.5855326652526855, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.028561925515532494, + "learning_rate": 0.0005705877537670184, + "loss": 4.637331008911133, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.026389246806502342, + "learning_rate": 0.0005704719784065846, + "loss": 4.598666667938232, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.02484819106757641, + "learning_rate": 0.0005703559875488088, + "loss": 4.601729393005371, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.02522977627813816, + "learning_rate": 0.0005702397812871429, + "loss": 4.609833717346191, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02589508146047592, + "learning_rate": 0.0005701233597152121, + "loss": 4.574592590332031, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.023374728858470917, + "learning_rate": 0.0005700067229268154, + "loss": 4.61362361907959, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.02465851791203022, + "learning_rate": 0.0005698898710159245, + "loss": 4.639899253845215, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.02258715033531189, + "learning_rate": 0.0005697728040766852, + "loss": 4.5656538009643555, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.018009789288043976, + "learning_rate": 0.0005696555222034162, + "loss": 4.61449670791626, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02121281623840332, + "learning_rate": 0.0005695380254906094, + "loss": 4.553830623626709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.018078701570630074, + "learning_rate": 0.0005694203140329296, + "loss": 4.534579277038574, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.0169303547590971, + "learning_rate": 0.000569302387925215, + "loss": 4.533995628356934, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.016318142414093018, + "learning_rate": 0.0005691842472624764, + "loss": 4.544438362121582, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.016664575785398483, + "learning_rate": 0.0005690658921398977, + "loss": 4.5913286209106445, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.01641504280269146, + "learning_rate": 0.0005689473226528354, + "loss": 4.520743370056152, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.017012735828757286, + "learning_rate": 0.0005688285388968187, + "loss": 4.517453193664551, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01672312617301941, + "learning_rate": 0.0005687095409675499, + "loss": 4.5235915184021, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.016893865540623665, + "learning_rate": 0.000568590328960903, + "loss": 4.543200492858887, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.017530135810375214, + "learning_rate": 0.0005684709029729253, + "loss": 4.54912805557251, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.018698599189519882, + "learning_rate": 0.0005683512630998361, + "loss": 4.522650718688965, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01984071172773838, + "learning_rate": 0.000568231409438027, + "loss": 4.546173095703125, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.023218289017677307, + "learning_rate": 0.0005681113420840619, + "loss": 4.5305094718933105, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.028968270868062973, + "learning_rate": 0.000567991061134677, + "loss": 4.521440505981445, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.032370273023843765, + "learning_rate": 0.0005678705666867805, + "loss": 4.54168701171875, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.033967748284339905, + "learning_rate": 0.0005677498588374524, + "loss": 4.538717269897461, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.028757929801940918, + "learning_rate": 0.0005676289376839452, + "loss": 4.517228603363037, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.029757825657725334, + "learning_rate": 0.0005675078033236827, + "loss": 4.510310173034668, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.03581003099679947, + "learning_rate": 0.0005673864558542605, + "loss": 4.506901264190674, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.03483985736966133, + "learning_rate": 0.0005672648953734462, + "loss": 4.549284934997559, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.031016027554869652, + "learning_rate": 0.000567143121979179, + "loss": 4.530084609985352, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.029373526573181152, + "learning_rate": 0.0005670211357695693, + "loss": 4.492218971252441, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.031661372631788254, + "learning_rate": 0.0005668989368428994, + "loss": 4.502824783325195, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.027251387014985085, + "learning_rate": 0.0005667765252976227, + "loss": 4.49193811416626, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02263176068663597, + "learning_rate": 0.0005666539012323639, + "loss": 4.512019157409668, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.02355228364467621, + "learning_rate": 0.0005665310647459189, + "loss": 4.508917331695557, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.023297371342778206, + "learning_rate": 0.0005664080159372551, + "loss": 4.497771263122559, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.019364114850759506, + "learning_rate": 0.0005662847549055107, + "loss": 4.51470947265625, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.023899901658296585, + "learning_rate": 0.0005661612817499947, + "loss": 4.48771858215332, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.02330983616411686, + "learning_rate": 0.0005660375965701872, + "loss": 4.478496551513672, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.027252234518527985, + "learning_rate": 0.0005659136994657392, + "loss": 4.497672080993652, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.030886363238096237, + "learning_rate": 0.0005657895905364723, + "loss": 4.49789571762085, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02515559457242489, + "learning_rate": 0.0005656652698823788, + "loss": 4.490771293640137, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020107125863432884, + "learning_rate": 0.0005655407376036217, + "loss": 4.481879234313965, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.022076716646552086, + "learning_rate": 0.0005654159938005345, + "loss": 4.494697570800781, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.025525815784931183, + "learning_rate": 0.0005652910385736207, + "loss": 4.517453193664551, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.03005894646048546, + "learning_rate": 0.0005651658720235548, + "loss": 4.453815460205078, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.030178502202033997, + "learning_rate": 0.0005650404942511812, + "loss": 4.493922233581543, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.029249394312500954, + "learning_rate": 0.0005649149053575141, + "loss": 4.453763008117676, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 4.607339382171631, + "eval_runtime": 57.9216, + "eval_samples_per_second": 42.16, + "eval_steps_per_second": 1.329, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.03362950682640076, + "learning_rate": 0.0005647891054437386, + "loss": 4.430700302124023, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.04223684221506119, + "learning_rate": 0.0005646630946112096, + "loss": 4.486915588378906, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.05192171037197113, + "learning_rate": 0.0005645368729614512, + "loss": 4.467541694641113, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.042525868862867355, + "learning_rate": 0.0005644104405961582, + "loss": 4.45252799987793, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.0406370609998703, + "learning_rate": 0.0005642837976171949, + "loss": 4.414548873901367, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.044476233422756195, + "learning_rate": 0.0005641569441265952, + "loss": 4.453492164611816, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.04687326401472092, + "learning_rate": 0.0005640298802265626, + "loss": 4.477725028991699, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.04011611267924309, + "learning_rate": 0.0005639026060194704, + "loss": 4.423356533050537, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.04971884936094284, + "learning_rate": 0.0005637751216078609, + "loss": 4.4853515625, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.05863234028220177, + "learning_rate": 0.0005636474270944461, + "loss": 4.466001033782959, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.057710036635398865, + "learning_rate": 0.0005635195225821072, + "loss": 4.466723442077637, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.05155074968934059, + "learning_rate": 0.0005633914081738945, + "loss": 4.4823198318481445, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.04535634070634842, + "learning_rate": 0.0005632630839730275, + "loss": 4.471219062805176, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.045147854834795, + "learning_rate": 0.0005631345500828946, + "loss": 4.476888179779053, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.03730619326233864, + "learning_rate": 0.0005630058066070533, + "loss": 4.428932189941406, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.034599728882312775, + "learning_rate": 0.0005628768536492299, + "loss": 4.47627592086792, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.032904356718063354, + "learning_rate": 0.0005627476913133193, + "loss": 4.422671794891357, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.03356337174773216, + "learning_rate": 0.0005626183197033856, + "loss": 4.456406593322754, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.035042595118284225, + "learning_rate": 0.0005624887389236609, + "loss": 4.371820449829102, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.03269032761454582, + "learning_rate": 0.0005623589490785462, + "loss": 4.402605056762695, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.028261031955480576, + "learning_rate": 0.0005622289502726108, + "loss": 4.433136940002441, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03301700949668884, + "learning_rate": 0.0005620987426105925, + "loss": 4.388359546661377, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.031664177775382996, + "learning_rate": 0.000561968326197397, + "loss": 4.41859245300293, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.026281699538230896, + "learning_rate": 0.0005618377011380988, + "loss": 4.406128883361816, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.023097911849617958, + "learning_rate": 0.0005617068675379398, + "loss": 4.389717102050781, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.022349286824464798, + "learning_rate": 0.0005615758255023306, + "loss": 4.408173561096191, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.021958226338028908, + "learning_rate": 0.0005614445751368491, + "loss": 4.434830188751221, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024551713839173317, + "learning_rate": 0.0005613131165472415, + "loss": 4.386944770812988, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.028048967942595482, + "learning_rate": 0.0005611814498394216, + "loss": 4.402279376983643, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02733907848596573, + "learning_rate": 0.0005610495751194708, + "loss": 4.403069496154785, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.020858559757471085, + "learning_rate": 0.0005609174924936384, + "loss": 4.406920909881592, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.021491149440407753, + "learning_rate": 0.0005607852020683409, + "loss": 4.385799407958984, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.021495329216122627, + "learning_rate": 0.0005606527039501621, + "loss": 4.4252495765686035, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.019708355888724327, + "learning_rate": 0.0005605199982458535, + "loss": 4.388864994049072, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.01886274851858616, + "learning_rate": 0.0005603870850623338, + "loss": 4.335962295532227, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.01955341547727585, + "learning_rate": 0.0005602539645066884, + "loss": 4.373560905456543, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.018962012603878975, + "learning_rate": 0.0005601206366861706, + "loss": 4.380022048950195, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.016663335263729095, + "learning_rate": 0.0005599871017081999, + "loss": 4.421034336090088, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01676873490214348, + "learning_rate": 0.0005598533596803631, + "loss": 4.382775783538818, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.014379126951098442, + "learning_rate": 0.0005597194107104137, + "loss": 4.41373872756958, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.015897583216428757, + "learning_rate": 0.000559585254906272, + "loss": 4.377284049987793, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.016513120383024216, + "learning_rate": 0.000559450892376025, + "loss": 4.41200065612793, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.016625918447971344, + "learning_rate": 0.000559316323227926, + "loss": 4.381187438964844, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.016028188169002533, + "learning_rate": 0.000559181547570395, + "loss": 4.353909015655518, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01682914048433304, + "learning_rate": 0.0005590465655120183, + "loss": 4.382326126098633, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.017078889533877373, + "learning_rate": 0.0005589113771615486, + "loss": 4.334949493408203, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.018182938918471336, + "learning_rate": 0.0005587759826279046, + "loss": 4.348130226135254, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01940886303782463, + "learning_rate": 0.0005586403820201713, + "loss": 4.339404106140137, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.020644692704081535, + "learning_rate": 0.0005585045754475995, + "loss": 4.358404159545898, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02433430217206478, + "learning_rate": 0.0005583685630196064, + "loss": 4.3660502433776855, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.027824537828564644, + "learning_rate": 0.0005582323448457744, + "loss": 4.34050178527832, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.02973582223057747, + "learning_rate": 0.0005580959210358523, + "loss": 4.348322868347168, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.030396128073334694, + "learning_rate": 0.0005579592916997542, + "loss": 4.356224060058594, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.029760530218482018, + "learning_rate": 0.0005578224569475599, + "loss": 4.357470989227295, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.02855013869702816, + "learning_rate": 0.0005576854168895147, + "loss": 4.335353374481201, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03376635164022446, + "learning_rate": 0.0005575481716360292, + "loss": 4.331794738769531, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.034431129693984985, + "learning_rate": 0.0005574107212976796, + "loss": 4.336880207061768, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.03524135425686836, + "learning_rate": 0.000557273065985207, + "loss": 4.31537389755249, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.03152058273553848, + "learning_rate": 0.0005571352058095179, + "loss": 4.37662410736084, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02878640592098236, + "learning_rate": 0.0005569971408816838, + "loss": 4.292388916015625, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.029663654044270515, + "learning_rate": 0.0005568588713129409, + "loss": 4.344346523284912, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.027895759791135788, + "learning_rate": 0.0005567203972146906, + "loss": 4.365072727203369, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.026465769857168198, + "learning_rate": 0.000556581718698499, + "loss": 4.334959506988525, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.03191852569580078, + "learning_rate": 0.000556442835876097, + "loss": 4.361069679260254, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.03535444661974907, + "learning_rate": 0.0005563037488593799, + "loss": 4.349765777587891, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03451581299304962, + "learning_rate": 0.0005561644577604074, + "loss": 4.343142509460449, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029810845851898193, + "learning_rate": 0.000556024962691404, + "loss": 4.307485580444336, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.027523217722773552, + "learning_rate": 0.0005558852637647584, + "loss": 4.366406440734863, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03043208085000515, + "learning_rate": 0.0005557453610930235, + "loss": 4.356929779052734, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.03058922104537487, + "learning_rate": 0.0005556052547889161, + "loss": 4.348942279815674, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.032052166759967804, + "learning_rate": 0.0005554649449653177, + "loss": 4.3496809005737305, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.03170694038271904, + "learning_rate": 0.0005553244317352731, + "loss": 4.315959930419922, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 4.4757890701293945, + "eval_runtime": 57.7451, + "eval_samples_per_second": 42.289, + "eval_steps_per_second": 1.333, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.029025450348854065, + "learning_rate": 0.0005551837152119915, + "loss": 4.288647174835205, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.029403962194919586, + "learning_rate": 0.0005550427955088455, + "loss": 4.288555145263672, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.03265019878745079, + "learning_rate": 0.0005549016727393715, + "loss": 4.292625904083252, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.0387202650308609, + "learning_rate": 0.0005547603470172697, + "loss": 4.238926887512207, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.04554206505417824, + "learning_rate": 0.0005546188184564036, + "loss": 4.29966402053833, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.036231476813554764, + "learning_rate": 0.0005544770871708004, + "loss": 4.304398536682129, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.038703009486198425, + "learning_rate": 0.0005543351532746501, + "loss": 4.2945356369018555, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03224619850516319, + "learning_rate": 0.0005541930168823067, + "loss": 4.261351585388184, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.029722856357693672, + "learning_rate": 0.0005540506781082864, + "loss": 4.303143501281738, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.02764045260846615, + "learning_rate": 0.0005539081370672695, + "loss": 4.243679523468018, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.0312200877815485, + "learning_rate": 0.0005537653938740984, + "loss": 4.279904365539551, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.027349065989255905, + "learning_rate": 0.0005536224486437788, + "loss": 4.270228385925293, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.030385233461856842, + "learning_rate": 0.0005534793014914793, + "loss": 4.301590442657471, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.03292274847626686, + "learning_rate": 0.0005533359525325307, + "loss": 4.301828384399414, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.03265977278351784, + "learning_rate": 0.0005531924018824266, + "loss": 4.286200523376465, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.03177836537361145, + "learning_rate": 0.0005530486496568236, + "loss": 4.296319007873535, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.028879934921860695, + "learning_rate": 0.0005529046959715397, + "loss": 4.261706829071045, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.02855812944471836, + "learning_rate": 0.0005527605409425562, + "loss": 4.273903846740723, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03037901036441326, + "learning_rate": 0.0005526161846860161, + "loss": 4.248080253601074, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.029279105365276337, + "learning_rate": 0.0005524716273182245, + "loss": 4.281311988830566, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.02592580020427704, + "learning_rate": 0.0005523268689556489, + "loss": 4.204848289489746, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.021899603307247162, + "learning_rate": 0.0005521819097149183, + "loss": 4.232453346252441, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.019704442471265793, + "learning_rate": 0.0005520367497128238, + "loss": 4.253462314605713, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.019550519064068794, + "learning_rate": 0.0005518913890663182, + "loss": 4.307737350463867, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.020320113748311996, + "learning_rate": 0.0005517458278925161, + "loss": 4.257411956787109, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02118559181690216, + "learning_rate": 0.0005516000663086932, + "loss": 4.262089729309082, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.02050616219639778, + "learning_rate": 0.0005514541044322872, + "loss": 4.2490644454956055, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019793255254626274, + "learning_rate": 0.0005513079423808967, + "loss": 4.267400741577148, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.020760254934430122, + "learning_rate": 0.0005511615802722823, + "loss": 4.2532429695129395, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.019524799659848213, + "learning_rate": 0.0005510150182243647, + "loss": 4.267422199249268, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.020166637375950813, + "learning_rate": 0.0005508682563552269, + "loss": 4.214296817779541, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.01964394748210907, + "learning_rate": 0.0005507212947831119, + "loss": 4.252087116241455, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01690433733165264, + "learning_rate": 0.0005505741336264239, + "loss": 4.232336044311523, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.017682762816548347, + "learning_rate": 0.0005504267730037283, + "loss": 4.213962554931641, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01688133366405964, + "learning_rate": 0.0005502792130337508, + "loss": 4.199461460113525, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.019166061654686928, + "learning_rate": 0.0005501314538353776, + "loss": 4.228936672210693, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.024065518751740456, + "learning_rate": 0.000549983495527656, + "loss": 4.214227676391602, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02518288791179657, + "learning_rate": 0.000549835338229793, + "loss": 4.256412029266357, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.026093967258930206, + "learning_rate": 0.0005496869820611563, + "loss": 4.216564178466797, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.023775635287165642, + "learning_rate": 0.0005495384271412739, + "loss": 4.226430892944336, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.018810907378792763, + "learning_rate": 0.0005493896735898338, + "loss": 4.190013885498047, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.018974872305989265, + "learning_rate": 0.000549240721526684, + "loss": 4.246231555938721, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.02069833129644394, + "learning_rate": 0.0005490915710718325, + "loss": 4.198457717895508, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.02097306028008461, + "learning_rate": 0.0005489422223454471, + "loss": 4.212507247924805, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022165542468428612, + "learning_rate": 0.0005487926754678556, + "loss": 4.2043843269348145, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.019068919122219086, + "learning_rate": 0.0005486429305595448, + "loss": 4.197460651397705, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01701812446117401, + "learning_rate": 0.0005484929877411618, + "loss": 4.220056533813477, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01629726029932499, + "learning_rate": 0.0005483428471335127, + "loss": 4.214868545532227, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018661806359887123, + "learning_rate": 0.000548192508857563, + "loss": 4.227047920227051, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.02005891129374504, + "learning_rate": 0.0005480419730344377, + "loss": 4.197265625, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.02591128647327423, + "learning_rate": 0.0005478912397854208, + "loss": 4.211041450500488, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.032659754157066345, + "learning_rate": 0.0005477403092319552, + "loss": 4.230195045471191, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.04057808965444565, + "learning_rate": 0.0005475891814956429, + "loss": 4.223843097686768, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.04103454574942589, + "learning_rate": 0.0005474378566982449, + "loss": 4.197145462036133, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.03810729831457138, + "learning_rate": 0.0005472863349616809, + "loss": 4.237957000732422, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.028838301077485085, + "learning_rate": 0.0005471346164080292, + "loss": 4.206722259521484, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02436191774904728, + "learning_rate": 0.0005469827011595266, + "loss": 4.1983137130737305, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.022839387878775597, + "learning_rate": 0.0005468305893385686, + "loss": 4.20530366897583, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.02375260926783085, + "learning_rate": 0.0005466782810677088, + "loss": 4.213202476501465, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.02177741937339306, + "learning_rate": 0.0005465257764696593, + "loss": 4.200539588928223, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.02351468615233898, + "learning_rate": 0.0005463730756672903, + "loss": 4.198091506958008, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.023375794291496277, + "learning_rate": 0.0005462201787836301, + "loss": 4.193244934082031, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.02479805052280426, + "learning_rate": 0.0005460670859418651, + "loss": 4.2250871658325195, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.026894841343164444, + "learning_rate": 0.000545913797265339, + "loss": 4.241791248321533, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.030582403764128685, + "learning_rate": 0.0005457603128775541, + "loss": 4.203666687011719, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.032222259789705276, + "learning_rate": 0.0005456066329021699, + "loss": 4.209629535675049, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.03181060031056404, + "learning_rate": 0.0005454527574630037, + "loss": 4.200459957122803, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02687576413154602, + "learning_rate": 0.0005452986866840298, + "loss": 4.211819648742676, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.025988668203353882, + "learning_rate": 0.0005451444206893806, + "loss": 4.198040962219238, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.02937682718038559, + "learning_rate": 0.0005449899596033453, + "loss": 4.222935676574707, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.03043951839208603, + "learning_rate": 0.0005448353035503704, + "loss": 4.2031755447387695, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.030401621013879776, + "learning_rate": 0.0005446804526550596, + "loss": 4.198380470275879, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 4.368444919586182, + "eval_runtime": 58.3009, + "eval_samples_per_second": 41.886, + "eval_steps_per_second": 1.321, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.03461585193872452, + "learning_rate": 0.0005445254070421732, + "loss": 4.147369384765625, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.039084456861019135, + "learning_rate": 0.0005443701668366288, + "loss": 4.124129295349121, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.04021848738193512, + "learning_rate": 0.0005442147321635007, + "loss": 4.127362251281738, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.041000768542289734, + "learning_rate": 0.00054405910314802, + "loss": 4.1432061195373535, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.04008613899350166, + "learning_rate": 0.0005439032799155737, + "loss": 4.169931888580322, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.039394307881593704, + "learning_rate": 0.000543747262591706, + "loss": 4.155813217163086, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.04374868795275688, + "learning_rate": 0.0005435910513021172, + "loss": 4.159891128540039, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0443582646548748, + "learning_rate": 0.0005434346461726641, + "loss": 4.157362937927246, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.04601646214723587, + "learning_rate": 0.0005432780473293593, + "loss": 4.152097225189209, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.04327242821455002, + "learning_rate": 0.0005431212548983718, + "loss": 4.170074462890625, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.0413462370634079, + "learning_rate": 0.0005429642690060262, + "loss": 4.156710147857666, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.036855220794677734, + "learning_rate": 0.0005428070897788034, + "loss": 4.138053894042969, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.03781217709183693, + "learning_rate": 0.0005426497173433398, + "loss": 4.172306060791016, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.03886301815509796, + "learning_rate": 0.0005424921518264275, + "loss": 4.135544776916504, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.03730994835495949, + "learning_rate": 0.0005423343933550143, + "loss": 4.171654224395752, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.03271950036287308, + "learning_rate": 0.0005421764420562032, + "loss": 4.169906139373779, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.03292673081159592, + "learning_rate": 0.0005420182980572527, + "loss": 4.163660049438477, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.029366906732320786, + "learning_rate": 0.0005418599614855768, + "loss": 4.161055088043213, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.023805437609553337, + "learning_rate": 0.0005417014324687442, + "loss": 4.172390937805176, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.023178137838840485, + "learning_rate": 0.000541542711134479, + "loss": 4.146788597106934, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.02403450198471546, + "learning_rate": 0.0005413837976106599, + "loss": 4.124364852905273, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.025009717792272568, + "learning_rate": 0.0005412246920253208, + "loss": 4.12275505065918, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.023431289941072464, + "learning_rate": 0.0005410653945066503, + "loss": 4.134739398956299, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.023803815245628357, + "learning_rate": 0.0005409059051829914, + "loss": 4.130463123321533, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.021219994872808456, + "learning_rate": 0.0005407462241828417, + "loss": 4.124577522277832, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.01899666152894497, + "learning_rate": 0.0005405863516348535, + "loss": 4.167086124420166, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.01889493502676487, + "learning_rate": 0.000540426287667833, + "loss": 4.152744293212891, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.019546914845705032, + "learning_rate": 0.000540266032410741, + "loss": 4.122655868530273, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02085448056459427, + "learning_rate": 0.0005401055859926923, + "loss": 4.141949653625488, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.019959403201937675, + "learning_rate": 0.0005399449485429555, + "loss": 4.131907939910889, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.01885392889380455, + "learning_rate": 0.0005397841201909535, + "loss": 4.118953227996826, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.018216131255030632, + "learning_rate": 0.0005396231010662627, + "loss": 4.096459865570068, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.019385207444429398, + "learning_rate": 0.0005394618912986132, + "loss": 4.111050128936768, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.018676603212952614, + "learning_rate": 0.0005393004910178889, + "loss": 4.156215667724609, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.01876339502632618, + "learning_rate": 0.0005391389003541271, + "loss": 4.1107282638549805, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.01831623539328575, + "learning_rate": 0.0005389771194375185, + "loss": 4.154116630554199, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.018896661698818207, + "learning_rate": 0.0005388151483984071, + "loss": 4.093320846557617, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.019582601264119148, + "learning_rate": 0.0005386529873672898, + "loss": 4.106500625610352, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.020313601940870285, + "learning_rate": 0.0005384906364748172, + "loss": 4.071961879730225, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.020385954529047012, + "learning_rate": 0.000538328095851792, + "loss": 4.129006862640381, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02231123298406601, + "learning_rate": 0.0005381653656291707, + "loss": 4.067756652832031, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.022957179695367813, + "learning_rate": 0.0005380024459380619, + "loss": 4.088230133056641, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021438462659716606, + "learning_rate": 0.000537839336909727, + "loss": 4.128036022186279, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01893841102719307, + "learning_rate": 0.00053767603867558, + "loss": 4.155407905578613, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.018539676442742348, + "learning_rate": 0.0005375125513671874, + "loss": 4.101880073547363, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.01881185919046402, + "learning_rate": 0.0005373488751162681, + "loss": 4.118906021118164, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.021657396107912064, + "learning_rate": 0.0005371850100546929, + "loss": 4.091935157775879, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.023293187841773033, + "learning_rate": 0.000537020956314485, + "loss": 4.105803489685059, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.021306317299604416, + "learning_rate": 0.0005368567140278196, + "loss": 4.0610809326171875, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.022078553214669228, + "learning_rate": 0.0005366922833270235, + "loss": 4.09867525100708, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.02266664244234562, + "learning_rate": 0.0005365276643445757, + "loss": 4.1027326583862305, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.02091757394373417, + "learning_rate": 0.0005363628572131069, + "loss": 4.080266952514648, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.020374659448862076, + "learning_rate": 0.0005361978620653991, + "loss": 4.088479042053223, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.02294750325381756, + "learning_rate": 0.0005360326790343858, + "loss": 4.084364891052246, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.02337353676557541, + "learning_rate": 0.000535867308253152, + "loss": 4.125696182250977, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.022643696516752243, + "learning_rate": 0.0005357017498549341, + "loss": 4.097160339355469, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.021080605685710907, + "learning_rate": 0.0005355360039731196, + "loss": 4.102363109588623, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.02269427664577961, + "learning_rate": 0.0005353700707412466, + "loss": 4.089844703674316, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.02195413038134575, + "learning_rate": 0.0005352039502930048, + "loss": 4.132293701171875, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.01805616170167923, + "learning_rate": 0.0005350376427622343, + "loss": 4.0842814445495605, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.018048783764243126, + "learning_rate": 0.0005348711482829261, + "loss": 4.109591007232666, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.0199508648365736, + "learning_rate": 0.000534704466989222, + "loss": 4.0908074378967285, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.020186785608530045, + "learning_rate": 0.0005345375990154138, + "loss": 4.077959060668945, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.022526193410158157, + "learning_rate": 0.0005343705444959441, + "loss": 4.109979629516602, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.025224441662430763, + "learning_rate": 0.000534203303565406, + "loss": 4.047533988952637, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.027292724698781967, + "learning_rate": 0.000534035876358542, + "loss": 4.09566593170166, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.025669090449810028, + "learning_rate": 0.0005338682630102454, + "loss": 4.0606160163879395, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02791150100529194, + "learning_rate": 0.0005337004636555595, + "loss": 4.097081184387207, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.03294133022427559, + "learning_rate": 0.0005335324784296768, + "loss": 4.121442794799805, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.027322066947817802, + "learning_rate": 0.0005333643074679403, + "loss": 4.1188740730285645, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.024866372346878052, + "learning_rate": 0.000533195950905842, + "loss": 4.082042694091797, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.02901027724146843, + "learning_rate": 0.0005330274088790239, + "loss": 4.110199451446533, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 4.283970355987549, + "eval_runtime": 58.4003, + "eval_samples_per_second": 41.815, + "eval_steps_per_second": 1.318, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.028958842158317566, + "learning_rate": 0.0005328586815232773, + "loss": 4.040733337402344, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.03491180017590523, + "learning_rate": 0.0005326897689745428, + "loss": 4.0473432540893555, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.04025818780064583, + "learning_rate": 0.0005325206713689101, + "loss": 4.021976470947266, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03998899459838867, + "learning_rate": 0.0005323513888426181, + "loss": 4.04901123046875, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.042165111750364304, + "learning_rate": 0.0005321819215320546, + "loss": 4.050230979919434, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.03511597216129303, + "learning_rate": 0.0005320122695737565, + "loss": 4.0989990234375, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.04064979776740074, + "learning_rate": 0.0005318424331044094, + "loss": 4.051448822021484, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.047536592930555344, + "learning_rate": 0.0005316724122608473, + "loss": 4.064278602600098, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.05592669919133186, + "learning_rate": 0.0005315022071800527, + "loss": 4.045516014099121, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.04262633994221687, + "learning_rate": 0.0005313318179991574, + "loss": 4.060666084289551, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03997781500220299, + "learning_rate": 0.0005311612448554402, + "loss": 4.065315246582031, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.038152433931827545, + "learning_rate": 0.0005309904878863293, + "loss": 4.069721221923828, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.0367623008787632, + "learning_rate": 0.0005308195472294001, + "loss": 4.080986976623535, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.0358598455786705, + "learning_rate": 0.0005306484230223769, + "loss": 4.065290927886963, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.039188165217638016, + "learning_rate": 0.000530477115403131, + "loss": 4.024333477020264, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.036839909851551056, + "learning_rate": 0.0005303056245096817, + "loss": 4.067532539367676, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.03456985577940941, + "learning_rate": 0.0005301339504801967, + "loss": 4.048443794250488, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.039487943053245544, + "learning_rate": 0.0005299620934529902, + "loss": 4.0585737228393555, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.03415827825665474, + "learning_rate": 0.0005297900535665244, + "loss": 4.06104850769043, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.03209620714187622, + "learning_rate": 0.0005296178309594087, + "loss": 4.009165287017822, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.028947360813617706, + "learning_rate": 0.0005294454257704, + "loss": 4.067384243011475, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026836825534701347, + "learning_rate": 0.0005292728381384018, + "loss": 4.046419620513916, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024095216765999794, + "learning_rate": 0.0005291000682024647, + "loss": 4.019008636474609, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.020965851843357086, + "learning_rate": 0.0005289271161017867, + "loss": 4.020613670349121, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.02250855043530464, + "learning_rate": 0.0005287539819757119, + "loss": 4.031634330749512, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.021385207772254944, + "learning_rate": 0.0005285806659637315, + "loss": 4.086246013641357, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.021104589104652405, + "learning_rate": 0.000528407168205483, + "loss": 4.033970355987549, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.019584037363529205, + "learning_rate": 0.0005282334888407503, + "loss": 4.025118827819824, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.02107258513569832, + "learning_rate": 0.000528059628009464, + "loss": 4.020823955535889, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.020438389852643013, + "learning_rate": 0.0005278855858517003, + "loss": 4.016323566436768, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.01881488971412182, + "learning_rate": 0.0005277113625076822, + "loss": 3.9991841316223145, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.01763618178665638, + "learning_rate": 0.000527536958117778, + "loss": 4.0062150955200195, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.017308862879872322, + "learning_rate": 0.0005273623728225022, + "loss": 3.946723461151123, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01706845872104168, + "learning_rate": 0.0005271876067625153, + "loss": 4.056562900543213, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.018664294853806496, + "learning_rate": 0.0005270126600786228, + "loss": 4.024652481079102, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.018523404374718666, + "learning_rate": 0.0005268375329117762, + "loss": 3.962064743041992, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.017408935353159904, + "learning_rate": 0.0005266622254030724, + "loss": 4.008697509765625, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.01588570512831211, + "learning_rate": 0.0005264867376937534, + "loss": 4.011664390563965, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.015084310434758663, + "learning_rate": 0.0005263110699252065, + "loss": 3.996224880218506, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01637054979801178, + "learning_rate": 0.000526135222238964, + "loss": 4.015972137451172, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.01673802360892296, + "learning_rate": 0.0005259591947767032, + "loss": 3.975414276123047, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.016544019803404808, + "learning_rate": 0.0005257829876802464, + "loss": 4.012955188751221, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.017455460503697395, + "learning_rate": 0.0005256066010915603, + "loss": 4.010180473327637, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.017625460401177406, + "learning_rate": 0.0005254300351527566, + "loss": 4.001679420471191, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.01732795126736164, + "learning_rate": 0.000525253290006091, + "loss": 4.006107330322266, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.016654646024107933, + "learning_rate": 0.0005250763657939641, + "loss": 4.045180797576904, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.015085658989846706, + "learning_rate": 0.0005248992626589203, + "loss": 3.994246482849121, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.015112568624317646, + "learning_rate": 0.0005247219807436489, + "loss": 4.004963397979736, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.014882124960422516, + "learning_rate": 0.000524544520190982, + "loss": 4.029565811157227, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.016012471169233322, + "learning_rate": 0.000524366881143897, + "loss": 4.018208026885986, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.015761924907565117, + "learning_rate": 0.0005241890637455141, + "loss": 3.9911205768585205, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.017770355567336082, + "learning_rate": 0.0005240110681390978, + "loss": 4.028733253479004, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01959751546382904, + "learning_rate": 0.0005238328944680558, + "loss": 4.003322124481201, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.018713824450969696, + "learning_rate": 0.0005236545428759394, + "loss": 3.991718292236328, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.01992208883166313, + "learning_rate": 0.0005234760135064434, + "loss": 4.002507209777832, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.021671291440725327, + "learning_rate": 0.0005232973065034055, + "loss": 4.021344184875488, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024282798171043396, + "learning_rate": 0.0005231184220108069, + "loss": 3.9899158477783203, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.023294344544410706, + "learning_rate": 0.0005229393601727713, + "loss": 4.046627044677734, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.02029707282781601, + "learning_rate": 0.000522760121133566, + "loss": 4.022454261779785, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.02151239849627018, + "learning_rate": 0.0005225807050376004, + "loss": 4.005476951599121, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.02320881187915802, + "learning_rate": 0.0005224011120294269, + "loss": 4.012726783752441, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.026484191417694092, + "learning_rate": 0.0005222213422537401, + "loss": 3.9904391765594482, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.026279661804437637, + "learning_rate": 0.0005220413958553777, + "loss": 3.9738848209381104, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.024299729615449905, + "learning_rate": 0.0005218612729793188, + "loss": 4.003316402435303, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.02337595447897911, + "learning_rate": 0.0005216809737706856, + "loss": 3.954984188079834, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.02416428178548813, + "learning_rate": 0.0005215004983747415, + "loss": 4.0317063331604, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.022488107904791832, + "learning_rate": 0.0005213198469368925, + "loss": 3.9736647605895996, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.019700225442647934, + "learning_rate": 0.0005211390196026861, + "loss": 3.979867458343506, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.020382333546876907, + "learning_rate": 0.0005209580165178117, + "loss": 3.995251178741455, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.020149298012256622, + "learning_rate": 0.0005207768378281003, + "loss": 3.9987268447875977, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.019790509715676308, + "learning_rate": 0.0005205954836795241, + "loss": 3.9838991165161133, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.022096451371908188, + "learning_rate": 0.000520413954218197, + "loss": 3.995800256729126, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.214347839355469, + "eval_runtime": 58.8872, + "eval_samples_per_second": 41.469, + "eval_steps_per_second": 1.308, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02203402668237686, + "learning_rate": 0.0005202322495903739, + "loss": 3.915332078933716, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.02336675301194191, + "learning_rate": 0.0005200503699424509, + "loss": 3.914801597595215, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.027339637279510498, + "learning_rate": 0.0005198683154209654, + "loss": 3.9277780055999756, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.026872653514146805, + "learning_rate": 0.0005196860861725952, + "loss": 3.9160094261169434, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.026558255776762962, + "learning_rate": 0.0005195036823441592, + "loss": 3.9084300994873047, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.02644367143511772, + "learning_rate": 0.0005193211040826169, + "loss": 3.9328718185424805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.024677276611328125, + "learning_rate": 0.0005191383515350681, + "loss": 3.88993501663208, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.023194244131445885, + "learning_rate": 0.0005189554248487536, + "loss": 3.923567056655884, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.024653147906064987, + "learning_rate": 0.0005187723241710539, + "loss": 3.8740999698638916, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.026041202247142792, + "learning_rate": 0.0005185890496494899, + "loss": 3.9542148113250732, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.029586758464574814, + "learning_rate": 0.0005184056014317225, + "loss": 3.9241347312927246, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.0325787179172039, + "learning_rate": 0.0005182219796655528, + "loss": 3.9656593799591064, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.03183380514383316, + "learning_rate": 0.0005180381844989214, + "loss": 3.950648307800293, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.03434471786022186, + "learning_rate": 0.0005178542160799088, + "loss": 3.9590439796447754, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.030099432915449142, + "learning_rate": 0.0005176700745567353, + "loss": 3.9518113136291504, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.025950564071536064, + "learning_rate": 0.0005174857600777599, + "loss": 3.8908939361572266, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.029959935694932938, + "learning_rate": 0.0005173012727914816, + "loss": 3.945037364959717, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.029551763087511063, + "learning_rate": 0.0005171166128465389, + "loss": 3.923959732055664, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.03538523241877556, + "learning_rate": 0.0005169317803917083, + "loss": 3.9395527839660645, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.037924155592918396, + "learning_rate": 0.0005167467755759065, + "loss": 3.945115089416504, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.036773014813661575, + "learning_rate": 0.0005165615985481881, + "loss": 3.9484646320343018, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.04125838354229927, + "learning_rate": 0.0005163762494577474, + "loss": 4.000699043273926, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.044125135987997055, + "learning_rate": 0.0005161907284539165, + "loss": 3.9828028678894043, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.040853384882211685, + "learning_rate": 0.0005160050356861663, + "loss": 3.957505702972412, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.03501008450984955, + "learning_rate": 0.0005158191713041063, + "loss": 3.9360849857330322, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.04082556813955307, + "learning_rate": 0.0005156331354574838, + "loss": 3.95941424369812, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.045432962477207184, + "learning_rate": 0.0005154469282961847, + "loss": 3.9666922092437744, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.04048244655132294, + "learning_rate": 0.0005152605499702328, + "loss": 3.984555244445801, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.031016314402222633, + "learning_rate": 0.0005150740006297896, + "loss": 3.945204257965088, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.03210771828889847, + "learning_rate": 0.0005148872804251548, + "loss": 3.976186752319336, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.03130023553967476, + "learning_rate": 0.000514700389506765, + "loss": 3.915727138519287, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.031193193048238754, + "learning_rate": 0.0005145133280251954, + "loss": 3.961866855621338, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.027518954128026962, + "learning_rate": 0.0005143260961311578, + "loss": 3.9742956161499023, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.02751155197620392, + "learning_rate": 0.0005141386939755014, + "loss": 3.949662208557129, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.026033353060483932, + "learning_rate": 0.0005139511217092129, + "loss": 3.9327926635742188, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.02579621411859989, + "learning_rate": 0.000513763379483416, + "loss": 3.9334418773651123, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.023009559139609337, + "learning_rate": 0.0005135754674493709, + "loss": 3.969707489013672, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.02196447364985943, + "learning_rate": 0.0005133873857584753, + "loss": 3.9084434509277344, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.02206484042108059, + "learning_rate": 0.0005131991345622629, + "loss": 3.922117233276367, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.023360205814242363, + "learning_rate": 0.0005130107140124046, + "loss": 3.965158224105835, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02162669226527214, + "learning_rate": 0.0005128221242607072, + "loss": 3.9519729614257812, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.019719846546649933, + "learning_rate": 0.0005126333654591141, + "loss": 3.9195961952209473, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.018988260999321938, + "learning_rate": 0.0005124444377597049, + "loss": 3.937926769256592, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.019278401508927345, + "learning_rate": 0.0005122553413146951, + "loss": 3.9126808643341064, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019101005047559738, + "learning_rate": 0.0005120660762764366, + "loss": 3.8993749618530273, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.019156700000166893, + "learning_rate": 0.0005118766427974166, + "loss": 3.9531068801879883, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019944407045841217, + "learning_rate": 0.0005116870410302582, + "loss": 3.9645180702209473, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.01826174184679985, + "learning_rate": 0.0005114972711277202, + "loss": 3.948477268218994, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.0191368218511343, + "learning_rate": 0.0005113073332426968, + "loss": 3.944648027420044, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.017768343910574913, + "learning_rate": 0.0005111172275282174, + "loss": 3.9584474563598633, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.01955590397119522, + "learning_rate": 0.0005109269541374469, + "loss": 3.8936800956726074, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.02100536972284317, + "learning_rate": 0.000510736513223685, + "loss": 3.93919038772583, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.020873935893177986, + "learning_rate": 0.0005105459049403665, + "loss": 3.929244041442871, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.021876046434044838, + "learning_rate": 0.0005103551294410614, + "loss": 3.9264731407165527, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.021726684644818306, + "learning_rate": 0.0005101641868794736, + "loss": 3.923135280609131, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.019308188930153847, + "learning_rate": 0.0005099730774094423, + "loss": 3.9062135219573975, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01917017623782158, + "learning_rate": 0.0005097818011849408, + "loss": 3.898810386657715, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017933931201696396, + "learning_rate": 0.0005095903583600772, + "loss": 3.8899734020233154, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.015131795778870583, + "learning_rate": 0.0005093987490890933, + "loss": 3.9452900886535645, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.017190825194120407, + "learning_rate": 0.0005092069735263651, + "loss": 3.9124746322631836, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.01618487387895584, + "learning_rate": 0.000509015031826403, + "loss": 3.91654109954834, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.014920198358595371, + "learning_rate": 0.0005088229241438509, + "loss": 3.9144535064697266, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01615205965936184, + "learning_rate": 0.0005086306506334864, + "loss": 3.917003631591797, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.015370192006230354, + "learning_rate": 0.0005084382114502208, + "loss": 3.8910603523254395, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.015290996059775352, + "learning_rate": 0.0005082456067490988, + "loss": 3.9122695922851562, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.015913106501102448, + "learning_rate": 0.0005080528366852987, + "loss": 3.9304556846618652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.01639600843191147, + "learning_rate": 0.0005078599014141318, + "loss": 3.8974502086639404, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.017010601237416267, + "learning_rate": 0.0005076668010910425, + "loss": 3.9723148345947266, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.017778996378183365, + "learning_rate": 0.0005074735358716083, + "loss": 3.8905391693115234, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.020029611885547638, + "learning_rate": 0.0005072801059115394, + "loss": 3.934976100921631, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.01931857503950596, + "learning_rate": 0.000507086511366679, + "loss": 3.9152774810791016, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019750485196709633, + "learning_rate": 0.0005068927523930027, + "loss": 3.877650499343872, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.161914825439453, + "eval_runtime": 58.5217, + "eval_samples_per_second": 41.728, + "eval_steps_per_second": 1.316, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02073049545288086, + "learning_rate": 0.0005066988291466184, + "loss": 3.8331246376037598, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02272389829158783, + "learning_rate": 0.0005065047417837669, + "loss": 3.8847155570983887, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02139078453183174, + "learning_rate": 0.0005063104904608206, + "loss": 3.8276169300079346, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02043229155242443, + "learning_rate": 0.0005061160753342843, + "loss": 3.8804001808166504, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.02345564216375351, + "learning_rate": 0.0005059214965607948, + "loss": 3.851609468460083, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.023448185995221138, + "learning_rate": 0.0005057267542971207, + "loss": 3.8446507453918457, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.024826066568493843, + "learning_rate": 0.0005055318487001624, + "loss": 3.86712384223938, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.02491980977356434, + "learning_rate": 0.0005053367799269518, + "loss": 3.8374528884887695, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.023002834990620613, + "learning_rate": 0.0005051415481346522, + "loss": 3.783214569091797, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.02123952843248844, + "learning_rate": 0.0005049461534805585, + "loss": 3.8463428020477295, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.01981225423514843, + "learning_rate": 0.0005047505961220964, + "loss": 3.8307082653045654, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01915367692708969, + "learning_rate": 0.0005045548762168231, + "loss": 3.8190226554870605, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.01890859752893448, + "learning_rate": 0.0005043589939224266, + "loss": 3.8428220748901367, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.019150376319885254, + "learning_rate": 0.0005041629493967257, + "loss": 3.830247402191162, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.020099453628063202, + "learning_rate": 0.0005039667427976697, + "loss": 3.838806629180908, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.021136781200766563, + "learning_rate": 0.0005037703742833391, + "loss": 3.8344273567199707, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.022857150062918663, + "learning_rate": 0.0005035738440119442, + "loss": 3.8317880630493164, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.024891739711165428, + "learning_rate": 0.0005033771521418261, + "loss": 3.837855577468872, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.02595464698970318, + "learning_rate": 0.0005031802988314557, + "loss": 3.846050262451172, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.02722409926354885, + "learning_rate": 0.0005029832842394342, + "loss": 3.8164305686950684, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.024486860260367393, + "learning_rate": 0.0005027861085244927, + "loss": 3.8099875450134277, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.021551169455051422, + "learning_rate": 0.0005025887718454922, + "loss": 3.825925827026367, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.02121506817638874, + "learning_rate": 0.0005023912743614232, + "loss": 3.870727777481079, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.01948230341076851, + "learning_rate": 0.0005021936162314058, + "loss": 3.8035314083099365, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.01896044984459877, + "learning_rate": 0.0005019957976146898, + "loss": 3.84665584564209, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.020620400086045265, + "learning_rate": 0.0005017978186706538, + "loss": 3.8338310718536377, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.020855262875556946, + "learning_rate": 0.0005015996795588062, + "loss": 3.8781003952026367, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.02153870463371277, + "learning_rate": 0.0005014013804387838, + "loss": 3.8783490657806396, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.023751024156808853, + "learning_rate": 0.0005012029214703529, + "loss": 3.846198558807373, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.023987066000699997, + "learning_rate": 0.000501004302813408, + "loss": 3.834296226501465, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.023626182228326797, + "learning_rate": 0.0005008055246279727, + "loss": 3.8591227531433105, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.025707192718982697, + "learning_rate": 0.000500606587074199, + "loss": 3.8766489028930664, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.027723029255867004, + "learning_rate": 0.0005004074903123674, + "loss": 3.88264799118042, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.0278895553201437, + "learning_rate": 0.0005002082345028864, + "loss": 3.8932995796203613, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.025622662156820297, + "learning_rate": 0.0005000088198062929, + "loss": 3.8734612464904785, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.02639615908265114, + "learning_rate": 0.0004998092463832516, + "loss": 3.786813735961914, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.026765599846839905, + "learning_rate": 0.0004996095143945554, + "loss": 3.853966474533081, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02823319099843502, + "learning_rate": 0.0004994096240011246, + "loss": 3.8321340084075928, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.023394955322146416, + "learning_rate": 0.0004992095753640072, + "loss": 3.8527331352233887, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.023874662816524506, + "learning_rate": 0.0004990093686443791, + "loss": 3.8822879791259766, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.02301311306655407, + "learning_rate": 0.000498809004003543, + "loss": 3.846186876296997, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.022630592808127403, + "learning_rate": 0.0004986084816029291, + "loss": 3.867992639541626, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.02424517087638378, + "learning_rate": 0.0004984078016040946, + "loss": 3.867293357849121, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.024116624146699905, + "learning_rate": 0.000498206964168724, + "loss": 3.846282958984375, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.02187768742442131, + "learning_rate": 0.000498005969458628, + "loss": 3.8429484367370605, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.01879809983074665, + "learning_rate": 0.0004978048176357447, + "loss": 3.863215446472168, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.020426234230399132, + "learning_rate": 0.0004976035088621385, + "loss": 3.825230121612549, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01992812193930149, + "learning_rate": 0.0004974020433000003, + "loss": 3.851665496826172, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.020693328231573105, + "learning_rate": 0.000497200421111647, + "loss": 3.906999111175537, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.021137919276952744, + "learning_rate": 0.0004969986424595221, + "loss": 3.8369035720825195, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.023285407572984695, + "learning_rate": 0.0004967967075061949, + "loss": 3.8658390045166016, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.0219135619699955, + "learning_rate": 0.0004965946164143609, + "loss": 3.873072862625122, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.020101983100175858, + "learning_rate": 0.0004963923693468413, + "loss": 3.847805976867676, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.01875176839530468, + "learning_rate": 0.0004961899664665827, + "loss": 3.7976622581481934, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01913360133767128, + "learning_rate": 0.0004959874079366576, + "loss": 3.853851318359375, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.016863619908690453, + "learning_rate": 0.0004957846939202637, + "loss": 3.877378225326538, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.017802366986870766, + "learning_rate": 0.000495581824580724, + "loss": 3.861356019973755, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.01981814205646515, + "learning_rate": 0.0004953788000814866, + "loss": 3.88613224029541, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.017508897930383682, + "learning_rate": 0.000495175620586125, + "loss": 3.8202030658721924, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.019240548834204674, + "learning_rate": 0.0004949722862583368, + "loss": 3.853827714920044, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.022305598482489586, + "learning_rate": 0.000494768797261945, + "loss": 3.885148763656616, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.02531025931239128, + "learning_rate": 0.000494565153760897, + "loss": 3.8412041664123535, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.024474849924445152, + "learning_rate": 0.0004943613559192648, + "loss": 3.865644931793213, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.020430924370884895, + "learning_rate": 0.0004941574039012444, + "loss": 3.8649442195892334, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.01906431093811989, + "learning_rate": 0.0004939532978711566, + "loss": 3.8462204933166504, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.021118057891726494, + "learning_rate": 0.0004937490379934456, + "loss": 3.8652522563934326, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.020337693393230438, + "learning_rate": 0.00049354462443268, + "loss": 3.8506221771240234, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.019090797752141953, + "learning_rate": 0.0004933400573535521, + "loss": 3.881237506866455, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.020973797887563705, + "learning_rate": 0.000493135336920878, + "loss": 3.821274757385254, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.021984362974762917, + "learning_rate": 0.0004929304632995973, + "loss": 3.8793835639953613, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.0215684212744236, + "learning_rate": 0.0004927254366547727, + "loss": 3.8259663581848145, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.021673865616321564, + "learning_rate": 0.0004925202571515908, + "loss": 3.8505711555480957, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.123871326446533, + "eval_runtime": 58.8357, + "eval_samples_per_second": 41.505, + "eval_steps_per_second": 1.309, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.0203359667211771, + "learning_rate": 0.0004923149249553608, + "loss": 3.785353660583496, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.023197868838906288, + "learning_rate": 0.0004921094402315152, + "loss": 3.7479515075683594, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.02314298041164875, + "learning_rate": 0.0004919038031456093, + "loss": 3.7697315216064453, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.0256908368319273, + "learning_rate": 0.0004916980138633213, + "loss": 3.79056978225708, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.026193121448159218, + "learning_rate": 0.0004914920725504518, + "loss": 3.7722532749176025, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.027120303362607956, + "learning_rate": 0.000491285979372924, + "loss": 3.7991981506347656, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.0266975536942482, + "learning_rate": 0.0004910797344967834, + "loss": 3.778301477432251, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.02806396596133709, + "learning_rate": 0.000490873338088198, + "loss": 3.7578704357147217, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.02893839031457901, + "learning_rate": 0.0004906667903134573, + "loss": 3.7401275634765625, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.031170370057225227, + "learning_rate": 0.0004904600913389735, + "loss": 3.780686616897583, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.03364266827702522, + "learning_rate": 0.0004902532413312799, + "loss": 3.794442653656006, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.030121006071567535, + "learning_rate": 0.0004900462404570321, + "loss": 3.779569149017334, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.03296204283833504, + "learning_rate": 0.0004898390888830067, + "loss": 3.7960336208343506, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.03381749242544174, + "learning_rate": 0.0004896317867761022, + "loss": 3.7959280014038086, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.035383883863687515, + "learning_rate": 0.000489424334303338, + "loss": 3.8008625507354736, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.03759761154651642, + "learning_rate": 0.0004892167316318548, + "loss": 3.7997028827667236, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.030651796609163284, + "learning_rate": 0.0004890089789289147, + "loss": 3.81191349029541, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.0262477844953537, + "learning_rate": 0.0004888010763618997, + "loss": 3.8046622276306152, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.030398324131965637, + "learning_rate": 0.0004885930240983136, + "loss": 3.7903637886047363, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.030515259131789207, + "learning_rate": 0.0004883848223057801, + "loss": 3.8111231327056885, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.031686048954725266, + "learning_rate": 0.00048817647115204386, + "loss": 3.781445026397705, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.025968654081225395, + "learning_rate": 0.00048796797080496936, + "loss": 3.788196563720703, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.025340190157294273, + "learning_rate": 0.00048775932143254176, + "loss": 3.763331413269043, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.024738596752285957, + "learning_rate": 0.0004875505232028661, + "loss": 3.8005969524383545, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.021990075707435608, + "learning_rate": 0.0004873415762841672, + "loss": 3.795844078063965, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.02191784791648388, + "learning_rate": 0.00048713248084478986, + "loss": 3.8062384128570557, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.023678259924054146, + "learning_rate": 0.0004869232370531986, + "loss": 3.7760679721832275, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.024137988686561584, + "learning_rate": 0.0004867138450779773, + "loss": 3.7917492389678955, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.02673223614692688, + "learning_rate": 0.0004865043050878293, + "loss": 3.8006649017333984, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.02395169623196125, + "learning_rate": 0.0004862946172515771, + "loss": 3.7902212142944336, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.02410844899713993, + "learning_rate": 0.0004860847817381625, + "loss": 3.8073291778564453, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.028109831735491753, + "learning_rate": 0.0004858747987166463, + "loss": 3.80116605758667, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.025538964197039604, + "learning_rate": 0.00048566466835620774, + "loss": 3.791151523590088, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.02393815852701664, + "learning_rate": 0.0004854543908261453, + "loss": 3.7766106128692627, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.024659741669893265, + "learning_rate": 0.0004852439662958756, + "loss": 3.789135217666626, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.02668776921927929, + "learning_rate": 0.000485033394934934, + "loss": 3.797905445098877, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.0258641317486763, + "learning_rate": 0.000484822676912974, + "loss": 3.783323287963867, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.024846022948622704, + "learning_rate": 0.0004846118123997674, + "loss": 3.7976717948913574, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.024981874972581863, + "learning_rate": 0.00048440080156520387, + "loss": 3.818732738494873, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.02556881681084633, + "learning_rate": 0.0004841896445792909, + "loss": 3.7821896076202393, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.02558571845293045, + "learning_rate": 0.000483978341612154, + "loss": 3.8143680095672607, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.02361566200852394, + "learning_rate": 0.0004837668928340362, + "loss": 3.7710189819335938, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022570716217160225, + "learning_rate": 0.0004835552984152978, + "loss": 3.775564670562744, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.022478004917502403, + "learning_rate": 0.00048334355852641664, + "loss": 3.7920689582824707, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021241728216409683, + "learning_rate": 0.0004831316733379879, + "loss": 3.7857837677001953, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.018921511247754097, + "learning_rate": 0.0004829196430207235, + "loss": 3.7659268379211426, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01926037296652794, + "learning_rate": 0.00048270746774545237, + "loss": 3.7527129650115967, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.018743667751550674, + "learning_rate": 0.0004824951476831204, + "loss": 3.767991065979004, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.017273949459195137, + "learning_rate": 0.00048228268300479015, + "loss": 3.7834174633026123, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.017054244875907898, + "learning_rate": 0.0004820700738816403, + "loss": 3.7775752544403076, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017190931364893913, + "learning_rate": 0.0004818573204849664, + "loss": 3.7709648609161377, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01665034517645836, + "learning_rate": 0.00048164442298617987, + "loss": 3.7575321197509766, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.016932843253016472, + "learning_rate": 0.00048143138155680844, + "loss": 3.7922799587249756, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01702750287950039, + "learning_rate": 0.0004812181963684958, + "loss": 3.7729780673980713, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01766599342226982, + "learning_rate": 0.00048100486759300134, + "loss": 3.7876572608947754, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01785697042942047, + "learning_rate": 0.00048079139540220024, + "loss": 3.8131699562072754, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01830688677728176, + "learning_rate": 0.0004805777799680831, + "loss": 3.790778636932373, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017348438501358032, + "learning_rate": 0.0004803640214627561, + "loss": 3.7984330654144287, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017999377101659775, + "learning_rate": 0.0004801501200584406, + "loss": 3.7756147384643555, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.01801210641860962, + "learning_rate": 0.000479936075927473, + "loss": 3.789759874343872, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.018445398658514023, + "learning_rate": 0.000479721889242305, + "loss": 3.7683000564575195, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.017544370144605637, + "learning_rate": 0.0004795075601755027, + "loss": 3.7776038646698, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.019834233447909355, + "learning_rate": 0.00047929308889974737, + "loss": 3.7559571266174316, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01791626401245594, + "learning_rate": 0.0004790784755878346, + "loss": 3.7746729850769043, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.017638562247157097, + "learning_rate": 0.00047886372041267434, + "loss": 3.7933082580566406, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.018974050879478455, + "learning_rate": 0.0004786488235472909, + "loss": 3.795713424682617, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.018694303929805756, + "learning_rate": 0.0004784337851648231, + "loss": 3.7675955295562744, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.01788092777132988, + "learning_rate": 0.0004782186054385233, + "loss": 3.786083698272705, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.01902497000992298, + "learning_rate": 0.0004780032845417579, + "loss": 3.7976233959198, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.020558880642056465, + "learning_rate": 0.000477787822648007, + "loss": 3.7500152587890625, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.020093899220228195, + "learning_rate": 0.00047757221993086443, + "loss": 3.703448534011841, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.019833799451589584, + "learning_rate": 0.00047735647656403754, + "loss": 3.756059408187866, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.093987941741943, + "eval_runtime": 57.6894, + "eval_samples_per_second": 42.33, + "eval_steps_per_second": 1.335, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.022167526185512543, + "learning_rate": 0.0004771405927213467, + "loss": 3.696444511413574, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.027874166145920753, + "learning_rate": 0.0004769245685767255, + "loss": 3.713630199432373, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.030872784554958344, + "learning_rate": 0.00047670840430422113, + "loss": 3.678295612335205, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.03246859833598137, + "learning_rate": 0.0004764921000779928, + "loss": 3.7186388969421387, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.03667069226503372, + "learning_rate": 0.0004762756560723131, + "loss": 3.7457995414733887, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.036286015063524246, + "learning_rate": 0.0004760590724615671, + "loss": 3.6881775856018066, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.03413614630699158, + "learning_rate": 0.0004758423494202522, + "loss": 3.6999263763427734, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.03384552150964737, + "learning_rate": 0.0004756254871229782, + "loss": 3.6959803104400635, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.03227181360125542, + "learning_rate": 0.00047540848574446724, + "loss": 3.7182722091674805, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.036366235464811325, + "learning_rate": 0.0004751913454595535, + "loss": 3.7116851806640625, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.03573326766490936, + "learning_rate": 0.0004749740664431828, + "loss": 3.741865873336792, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.03482096269726753, + "learning_rate": 0.000474756648870413, + "loss": 3.7195520401000977, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.03232625126838684, + "learning_rate": 0.0004745390929164134, + "loss": 3.7443251609802246, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.03421664610505104, + "learning_rate": 0.0004743213987564651, + "loss": 3.6867361068725586, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.03585013002157211, + "learning_rate": 0.0004741035665659603, + "loss": 3.7492403984069824, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.029985329136252403, + "learning_rate": 0.00047388559652040236, + "loss": 3.705209493637085, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.02729589305818081, + "learning_rate": 0.0004736674887954059, + "loss": 3.711106538772583, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.029642216861248016, + "learning_rate": 0.0004734492435666963, + "loss": 3.7274727821350098, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.02816903032362461, + "learning_rate": 0.0004732308610101099, + "loss": 3.7181742191314697, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.027143625542521477, + "learning_rate": 0.00047301234130159346, + "loss": 3.719235897064209, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.025802744552493095, + "learning_rate": 0.00047279368461720445, + "loss": 3.771820545196533, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.026424532756209373, + "learning_rate": 0.0004725748911331106, + "loss": 3.7516651153564453, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.029677648097276688, + "learning_rate": 0.0004723559610255898, + "loss": 3.7058472633361816, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.031548649072647095, + "learning_rate": 0.0004721368944710302, + "loss": 3.756992816925049, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.027667395770549774, + "learning_rate": 0.00047191769164592974, + "loss": 3.742711067199707, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.027639545500278473, + "learning_rate": 0.000471698352726896, + "loss": 3.728883743286133, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.02950332500040531, + "learning_rate": 0.0004714788778906467, + "loss": 3.7160212993621826, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.026302607730031013, + "learning_rate": 0.0004712592673140084, + "loss": 3.735713481903076, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.022483112290501595, + "learning_rate": 0.00047103952117391764, + "loss": 3.702686309814453, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.021729422733187675, + "learning_rate": 0.00047081963964741986, + "loss": 3.7195816040039062, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.021159254014492035, + "learning_rate": 0.0004705996229116696, + "loss": 3.684321880340576, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.02233424223959446, + "learning_rate": 0.0004703794711439304, + "loss": 3.7233057022094727, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.02142048068344593, + "learning_rate": 0.0004701591845215744, + "loss": 3.7083680629730225, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.022331461310386658, + "learning_rate": 0.00046993876322208274, + "loss": 3.7408461570739746, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.020629866048693657, + "learning_rate": 0.0004697182074230448, + "loss": 3.759129047393799, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.02174685336649418, + "learning_rate": 0.0004694975173021586, + "loss": 3.72351336479187, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.021538110449910164, + "learning_rate": 0.0004692766930372299, + "loss": 3.716416597366333, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.020800625905394554, + "learning_rate": 0.00046905573480617285, + "loss": 3.726954936981201, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.02039588987827301, + "learning_rate": 0.00046883464278700966, + "loss": 3.7336390018463135, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.02107219770550728, + "learning_rate": 0.0004686134171578702, + "loss": 3.722311496734619, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.02068956382572651, + "learning_rate": 0.00046839205809699183, + "loss": 3.767139434814453, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.020190054550766945, + "learning_rate": 0.0004681705657827195, + "loss": 3.6960511207580566, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.018814411014318466, + "learning_rate": 0.0004679489403935058, + "loss": 3.70375919342041, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.017568519338965416, + "learning_rate": 0.00046772718210791005, + "loss": 3.756627321243286, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.018188195303082466, + "learning_rate": 0.00046750529110459903, + "loss": 3.6964683532714844, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.018925746902823448, + "learning_rate": 0.0004672832675623463, + "loss": 3.731172561645508, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.01690756157040596, + "learning_rate": 0.0004670611116600322, + "loss": 3.6955318450927734, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016300756484270096, + "learning_rate": 0.00046683882357664375, + "loss": 3.762632369995117, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.016405068337917328, + "learning_rate": 0.0004666164034912744, + "loss": 3.75130558013916, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.017687877640128136, + "learning_rate": 0.00046639385158312406, + "loss": 3.687530755996704, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.017524344846606255, + "learning_rate": 0.00046617116803149874, + "loss": 3.738790273666382, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.01714804768562317, + "learning_rate": 0.00046594835301581053, + "loss": 3.69827938079834, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018042324110865593, + "learning_rate": 0.00046572540671557745, + "loss": 3.71600341796875, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018648479133844376, + "learning_rate": 0.00046550232931042346, + "loss": 3.742591619491577, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.019198937341570854, + "learning_rate": 0.00046527912098007787, + "loss": 3.739820957183838, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.01796373911201954, + "learning_rate": 0.00046505578190437564, + "loss": 3.719025135040283, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01836620457470417, + "learning_rate": 0.0004648323122632569, + "loss": 3.741434097290039, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.02192627638578415, + "learning_rate": 0.0004646087122367673, + "loss": 3.7553882598876953, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.02147185243666172, + "learning_rate": 0.00046438498200505727, + "loss": 3.7705790996551514, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01855960488319397, + "learning_rate": 0.0004641611217483821, + "loss": 3.704925537109375, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.01925475336611271, + "learning_rate": 0.00046393713164710217, + "loss": 3.711491346359253, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019796503707766533, + "learning_rate": 0.00046371301188168204, + "loss": 3.7241992950439453, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.018511971458792686, + "learning_rate": 0.0004634887626326911, + "loss": 3.6785435676574707, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.018214019015431404, + "learning_rate": 0.00046326438408080293, + "loss": 3.7299962043762207, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.018912190571427345, + "learning_rate": 0.00046303987640679517, + "loss": 3.773698329925537, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.01939854398369789, + "learning_rate": 0.0004628152397915498, + "loss": 3.682321071624756, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.017859861254692078, + "learning_rate": 0.00046259047441605215, + "loss": 3.7254951000213623, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.019249707460403442, + "learning_rate": 0.0004623655804613919, + "loss": 3.7263433933258057, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.019221967086195946, + "learning_rate": 0.00046214055810876194, + "loss": 3.7229225635528564, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01871953159570694, + "learning_rate": 0.00046191540753945886, + "loss": 3.7248311042785645, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01747182011604309, + "learning_rate": 0.00046169012893488204, + "loss": 3.699913740158081, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.019630730152130127, + "learning_rate": 0.00046146472247653484, + "loss": 3.787783145904541, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.072091102600098, + "eval_runtime": 63.1589, + "eval_samples_per_second": 38.664, + "eval_steps_per_second": 1.219, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.022918934002518654, + "learning_rate": 0.0004612391883460228, + "loss": 3.666675567626953, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.0264838095754385, + "learning_rate": 0.00046101352672505493, + "loss": 3.6371259689331055, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.028463594615459442, + "learning_rate": 0.0004607877377954426, + "loss": 3.657130241394043, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.02900792844593525, + "learning_rate": 0.00046056182173909995, + "loss": 3.6365628242492676, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.028593607246875763, + "learning_rate": 0.0004603357787380435, + "loss": 3.647258996963501, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.02878790907561779, + "learning_rate": 0.0004601096089743919, + "loss": 3.682035446166992, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.02980183996260166, + "learning_rate": 0.0004598833126303661, + "loss": 3.623297691345215, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.03375725448131561, + "learning_rate": 0.00045965688988828884, + "loss": 3.6240830421447754, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.035045325756073, + "learning_rate": 0.00045943034093058507, + "loss": 3.6372249126434326, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.03519923985004425, + "learning_rate": 0.000459203665939781, + "loss": 3.6157641410827637, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.03468473628163338, + "learning_rate": 0.00045897686509850487, + "loss": 3.6553335189819336, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.034817393869161606, + "learning_rate": 0.0004587499385894856, + "loss": 3.653806686401367, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.03549059107899666, + "learning_rate": 0.000458522886595554, + "loss": 3.6730189323425293, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.028001410886645317, + "learning_rate": 0.0004582957092996418, + "loss": 3.662490129470825, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.02532520331442356, + "learning_rate": 0.0004580684068847817, + "loss": 3.629936456680298, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.028215918689966202, + "learning_rate": 0.0004578409795341069, + "loss": 3.6507444381713867, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.026823483407497406, + "learning_rate": 0.0004576134274308517, + "loss": 3.661811113357544, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.0281058382242918, + "learning_rate": 0.0004573857507583506, + "loss": 3.6805872917175293, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.02548503689467907, + "learning_rate": 0.00045715794970003874, + "loss": 3.6586413383483887, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.024795135483145714, + "learning_rate": 0.0004569300244394513, + "loss": 3.654989004135132, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.025501688942313194, + "learning_rate": 0.0004567019751602234, + "loss": 3.6915903091430664, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.02377370186150074, + "learning_rate": 0.00045647380204609034, + "loss": 3.660588264465332, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.022150425240397453, + "learning_rate": 0.00045624550528088704, + "loss": 3.64870023727417, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.02415974996984005, + "learning_rate": 0.00045601708504854803, + "loss": 3.6498191356658936, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.02140078693628311, + "learning_rate": 0.0004557885415331074, + "loss": 3.703489065170288, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.022121304646134377, + "learning_rate": 0.00045555987491869865, + "loss": 3.6606979370117188, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.02168990671634674, + "learning_rate": 0.00045533108538955416, + "loss": 3.6801178455352783, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.020228713750839233, + "learning_rate": 0.0004551021731300055, + "loss": 3.6707205772399902, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.018500352278351784, + "learning_rate": 0.0004548731383244834, + "loss": 3.647545576095581, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.020223816856741905, + "learning_rate": 0.0004546439811575169, + "loss": 3.644932985305786, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.019502121955156326, + "learning_rate": 0.00045441470181373375, + "loss": 3.62888240814209, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.01926770992577076, + "learning_rate": 0.0004541853004778603, + "loss": 3.675457715988159, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01888471283018589, + "learning_rate": 0.00045395577733472087, + "loss": 3.636434316635132, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.0201494712382555, + "learning_rate": 0.0004537261325692383, + "loss": 3.679147720336914, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.02178201824426651, + "learning_rate": 0.0004534963663664332, + "loss": 3.671905040740967, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.024073559790849686, + "learning_rate": 0.00045326647891142394, + "loss": 3.6701040267944336, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.02135559171438217, + "learning_rate": 0.0004530364703894268, + "loss": 3.648519992828369, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.019264448434114456, + "learning_rate": 0.0004528063409857554, + "loss": 3.6985023021698, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.01922295056283474, + "learning_rate": 0.0004525760908858209, + "loss": 3.651756763458252, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.01978331431746483, + "learning_rate": 0.00045234572027513174, + "loss": 3.679884910583496, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.01968599483370781, + "learning_rate": 0.00045211522933929317, + "loss": 3.6571812629699707, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.020741639658808708, + "learning_rate": 0.0004518846182640077, + "loss": 3.6505298614501953, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.01842017099261284, + "learning_rate": 0.0004516538872350744, + "loss": 3.6904940605163574, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.018004486337304115, + "learning_rate": 0.0004514230364383893, + "loss": 3.7282047271728516, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.018821008503437042, + "learning_rate": 0.0004511920660599444, + "loss": 3.668164014816284, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.018929854035377502, + "learning_rate": 0.0004509609762858287, + "loss": 3.6672980785369873, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.01979168877005577, + "learning_rate": 0.0004507297673022269, + "loss": 3.6684999465942383, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.01774384267628193, + "learning_rate": 0.0004504984392954199, + "loss": 3.6577749252319336, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01876000687479973, + "learning_rate": 0.00045026699245178444, + "loss": 3.6629252433776855, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.018197249621152878, + "learning_rate": 0.0004500354269577932, + "loss": 3.688983678817749, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.019496265798807144, + "learning_rate": 0.00044980374300001414, + "loss": 3.6602022647857666, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01825535111129284, + "learning_rate": 0.00044957194076511095, + "loss": 3.6889238357543945, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.017113137990236282, + "learning_rate": 0.0004493400204398426, + "loss": 3.684070587158203, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.017993126064538956, + "learning_rate": 0.00044910798221106283, + "loss": 3.675426721572876, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.018354708328843117, + "learning_rate": 0.00044887582626572105, + "loss": 3.675588607788086, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.016362832859158516, + "learning_rate": 0.0004486435527908606, + "loss": 3.6709303855895996, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01772182621061802, + "learning_rate": 0.0004484111619736205, + "loss": 3.6884102821350098, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.019271764904260635, + "learning_rate": 0.0004481786540012336, + "loss": 3.6468968391418457, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.019040528684854507, + "learning_rate": 0.00044794602906102747, + "loss": 3.690387010574341, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.019022610038518906, + "learning_rate": 0.00044771328734042367, + "loss": 3.689873218536377, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.018533391878008842, + "learning_rate": 0.00044748042902693815, + "loss": 3.6877284049987793, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.020825615152716637, + "learning_rate": 0.0004472474543081805, + "loss": 3.662257671356201, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.021880341693758965, + "learning_rate": 0.00044701436337185444, + "loss": 3.650700092315674, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.019974704831838608, + "learning_rate": 0.00044678115640575704, + "loss": 3.684177875518799, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.01903127320110798, + "learning_rate": 0.0004465478335977788, + "loss": 3.658148765563965, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.020999105647206306, + "learning_rate": 0.00044631439513590393, + "loss": 3.68747615814209, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.02138398587703705, + "learning_rate": 0.00044608084120820933, + "loss": 3.656355381011963, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.02216442860662937, + "learning_rate": 0.0004458471720028655, + "loss": 3.6860342025756836, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.019359121099114418, + "learning_rate": 0.0004456133877081352, + "loss": 3.654944658279419, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.020984770730137825, + "learning_rate": 0.0004453794885123744, + "loss": 3.6885194778442383, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.02423972077667713, + "learning_rate": 0.00044514547460403135, + "loss": 3.672712802886963, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.024544453248381615, + "learning_rate": 0.00044491134617164694, + "loss": 3.6484768390655518, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.060352325439453, + "eval_runtime": 57.9556, + "eval_samples_per_second": 42.136, + "eval_steps_per_second": 1.329, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.024072550237178802, + "learning_rate": 0.000444677103403854, + "loss": 3.5796468257904053, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.028701433911919594, + "learning_rate": 0.0004444427464893779, + "loss": 3.604318618774414, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.028119578957557678, + "learning_rate": 0.00044420827561703573, + "loss": 3.608759641647339, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.025335192680358887, + "learning_rate": 0.0004439736909757365, + "loss": 3.5760319232940674, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.0273929201066494, + "learning_rate": 0.0004437389927544805, + "loss": 3.559441089630127, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.026567714288830757, + "learning_rate": 0.0004435041811423603, + "loss": 3.626889228820801, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.027587885037064552, + "learning_rate": 0.0004432692563285591, + "loss": 3.6226272583007812, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.02935059368610382, + "learning_rate": 0.0004430342185023517, + "loss": 3.601921558380127, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.02982175536453724, + "learning_rate": 0.0004427990678531038, + "loss": 3.59737491607666, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.028570258989930153, + "learning_rate": 0.0004425638045702719, + "loss": 3.56337833404541, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.029890350997447968, + "learning_rate": 0.0004423284288434035, + "loss": 3.6294307708740234, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.03206814453005791, + "learning_rate": 0.00044209294086213657, + "loss": 3.590505599975586, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.03169345110654831, + "learning_rate": 0.00044185734081619957, + "loss": 3.581584930419922, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.03313758969306946, + "learning_rate": 0.000441621628895411, + "loss": 3.6088438034057617, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.03317603841423988, + "learning_rate": 0.00044138580528967986, + "loss": 3.6340456008911133, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.034738484770059586, + "learning_rate": 0.0004411498701890051, + "loss": 3.5814096927642822, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.036428093910217285, + "learning_rate": 0.0004409138237834751, + "loss": 3.5939526557922363, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.03293406963348389, + "learning_rate": 0.00044067766626326835, + "loss": 3.577897787094116, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.031030720099806786, + "learning_rate": 0.00044044139781865255, + "loss": 3.6077990531921387, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.03441236540675163, + "learning_rate": 0.00044020501863998516, + "loss": 3.591923713684082, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.03500965237617493, + "learning_rate": 0.00043996852891771236, + "loss": 3.6171927452087402, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.034443050622940063, + "learning_rate": 0.00043973192884236983, + "loss": 3.619724750518799, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.03435051813721657, + "learning_rate": 0.0004394952186045819, + "loss": 3.6443169116973877, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.03230702877044678, + "learning_rate": 0.00043925839839506167, + "loss": 3.5912671089172363, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.03232988342642784, + "learning_rate": 0.0004390214684046111, + "loss": 3.618985414505005, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.029298555105924606, + "learning_rate": 0.00043878442882412015, + "loss": 3.6166677474975586, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.026723450049757957, + "learning_rate": 0.00043854727984456746, + "loss": 3.61362886428833, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.028280874714255333, + "learning_rate": 0.0004383100216570197, + "loss": 3.6410951614379883, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.026700228452682495, + "learning_rate": 0.0004380726544526316, + "loss": 3.5804049968719482, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.023092713207006454, + "learning_rate": 0.00043783517842264534, + "loss": 3.595339298248291, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.024123897776007652, + "learning_rate": 0.0004375975937583914, + "loss": 3.647569417953491, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.02505572699010372, + "learning_rate": 0.0004373599006512871, + "loss": 3.6276652812957764, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.024523509666323662, + "learning_rate": 0.00043712209929283786, + "loss": 3.59525728225708, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.024351581931114197, + "learning_rate": 0.00043688418987463567, + "loss": 3.6081674098968506, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.02530614472925663, + "learning_rate": 0.00043664617258835995, + "loss": 3.6632542610168457, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.023438062518835068, + "learning_rate": 0.0004364080476257769, + "loss": 3.6345431804656982, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.021853657439351082, + "learning_rate": 0.0004361698151787396, + "loss": 3.598644256591797, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.022487422451376915, + "learning_rate": 0.0004359314754391874, + "loss": 3.5966577529907227, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.021353935822844505, + "learning_rate": 0.0004356930285991466, + "loss": 3.6328375339508057, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.021180452778935432, + "learning_rate": 0.00043545447485072926, + "loss": 3.6285812854766846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.020491832867264748, + "learning_rate": 0.000435215814386134, + "loss": 3.592944622039795, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.021603770554065704, + "learning_rate": 0.0004349770473976453, + "loss": 3.613507032394409, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.02146158367395401, + "learning_rate": 0.0004347381740776332, + "loss": 3.6409506797790527, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.02102433145046234, + "learning_rate": 0.0004344991946185538, + "loss": 3.6179330348968506, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.018684551119804382, + "learning_rate": 0.00043426010921294837, + "loss": 3.6385130882263184, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.018416227772831917, + "learning_rate": 0.0004340209180534438, + "loss": 3.642151355743408, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.01915144734084606, + "learning_rate": 0.000433781621332752, + "loss": 3.6089656352996826, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.018299665302038193, + "learning_rate": 0.0004335422192436703, + "loss": 3.6063547134399414, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.018962757661938667, + "learning_rate": 0.00043330271197908023, + "loss": 3.625917434692383, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.017437821254134178, + "learning_rate": 0.00043306309973194874, + "loss": 3.600015640258789, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.017913147807121277, + "learning_rate": 0.000432823382695327, + "loss": 3.6428744792938232, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.018343618139624596, + "learning_rate": 0.0004325835610623508, + "loss": 3.635901927947998, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.020372850820422173, + "learning_rate": 0.00043234363502623994, + "loss": 3.620181083679199, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.02037600241601467, + "learning_rate": 0.0004321036047802985, + "loss": 3.61129093170166, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.021827151998877525, + "learning_rate": 0.00043186347051791466, + "loss": 3.5999555587768555, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.021806657314300537, + "learning_rate": 0.0004316232324325602, + "loss": 3.642509937286377, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02050699107348919, + "learning_rate": 0.0004313828907177906, + "loss": 3.597583293914795, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.020436758175492287, + "learning_rate": 0.0004311424455672448, + "loss": 3.6079912185668945, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.019341666251420975, + "learning_rate": 0.0004309018971746451, + "loss": 3.6196017265319824, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.022348817437887192, + "learning_rate": 0.000430661245733797, + "loss": 3.619025468826294, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.019361121580004692, + "learning_rate": 0.00043042049143858914, + "loss": 3.657182455062866, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.019258935004472733, + "learning_rate": 0.0004301796344829928, + "loss": 3.6408963203430176, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.020622272044420242, + "learning_rate": 0.00042993867506106205, + "loss": 3.6247220039367676, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.019510755315423012, + "learning_rate": 0.0004296976133669336, + "loss": 3.6264584064483643, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.02078665979206562, + "learning_rate": 0.0004294564495948266, + "loss": 3.6364388465881348, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01791800744831562, + "learning_rate": 0.0004292151839390421, + "loss": 3.6029374599456787, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.01922667771577835, + "learning_rate": 0.0004289738165939636, + "loss": 3.582937717437744, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.018473893404006958, + "learning_rate": 0.00042873234775405654, + "loss": 3.6141819953918457, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01846298575401306, + "learning_rate": 0.00042849077761386784, + "loss": 3.638786554336548, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.01876973733305931, + "learning_rate": 0.0004282491063680263, + "loss": 3.617130756378174, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.017995499074459076, + "learning_rate": 0.00042800733421124207, + "loss": 3.645305633544922, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.02105429396033287, + "learning_rate": 0.00042776546133830646, + "loss": 3.6444876194000244, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.049062252044678, + "eval_runtime": 58.0113, + "eval_samples_per_second": 42.095, + "eval_steps_per_second": 1.327, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.02128119394183159, + "learning_rate": 0.00042752348794409227, + "loss": 3.524954080581665, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.028249500319361687, + "learning_rate": 0.00042728141422355287, + "loss": 3.5311479568481445, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.03145391866564751, + "learning_rate": 0.0004270392403717229, + "loss": 3.559330940246582, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.02994118444621563, + "learning_rate": 0.00042679696658371743, + "loss": 3.5257620811462402, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.027448246255517006, + "learning_rate": 0.00042655459305473195, + "loss": 3.5242764949798584, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.0296134352684021, + "learning_rate": 0.0004263121199800425, + "loss": 3.5317091941833496, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.029576804488897324, + "learning_rate": 0.0004260695475550054, + "loss": 3.5340428352355957, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.02941269613802433, + "learning_rate": 0.0004258268759750566, + "loss": 3.5372023582458496, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.03126693144440651, + "learning_rate": 0.0004255841054357124, + "loss": 3.5437135696411133, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.03335367515683174, + "learning_rate": 0.0004253412361325686, + "loss": 3.5598273277282715, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.03028946928679943, + "learning_rate": 0.0004250982682613006, + "loss": 3.5190839767456055, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.028629133477807045, + "learning_rate": 0.00042485520201766315, + "loss": 3.5138821601867676, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.027070026844739914, + "learning_rate": 0.0004246120375974905, + "loss": 3.530958414077759, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.028837760910391808, + "learning_rate": 0.0004243687751966957, + "loss": 3.5233962535858154, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.029884297400712967, + "learning_rate": 0.0004241254150112708, + "loss": 3.5639758110046387, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.024223247542977333, + "learning_rate": 0.000423881957237287, + "loss": 3.5555152893066406, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.026992836967110634, + "learning_rate": 0.0004236384020708935, + "loss": 3.540196418762207, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.02818330004811287, + "learning_rate": 0.0004233947497083185, + "loss": 3.5546071529388428, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.026401042938232422, + "learning_rate": 0.0004231510003458682, + "loss": 3.5556247234344482, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.026020489633083344, + "learning_rate": 0.0004229071541799272, + "loss": 3.568004608154297, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.02803782932460308, + "learning_rate": 0.0004226632114069578, + "loss": 3.5828235149383545, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.02737978659570217, + "learning_rate": 0.0004224191722235004, + "loss": 3.5642499923706055, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.02406371757388115, + "learning_rate": 0.00042217503682617286, + "loss": 3.5530993938446045, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.02230692468583584, + "learning_rate": 0.0004219308054116706, + "loss": 3.5573501586914062, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.021352406591176987, + "learning_rate": 0.0004216864781767667, + "loss": 3.5526974201202393, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.02227999083697796, + "learning_rate": 0.00042144205531831103, + "loss": 3.5866928100585938, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.024190247058868408, + "learning_rate": 0.0004211975370332308, + "loss": 3.517117977142334, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.023196619004011154, + "learning_rate": 0.00042095292351852976, + "loss": 3.4826502799987793, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.02388714998960495, + "learning_rate": 0.0004207082149712888, + "loss": 3.5588626861572266, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.024692339822649956, + "learning_rate": 0.0004204634115886651, + "loss": 3.54828143119812, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.023078206926584244, + "learning_rate": 0.0004202185135678924, + "loss": 3.5679030418395996, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.020665735006332397, + "learning_rate": 0.00041997352110628033, + "loss": 3.5500283241271973, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.02022322081029415, + "learning_rate": 0.00041972843440121527, + "loss": 3.592104911804199, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.021137656643986702, + "learning_rate": 0.00041948325365015894, + "loss": 3.583833694458008, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.021026931703090668, + "learning_rate": 0.0004192379790506491, + "loss": 3.556403875350952, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.020423648878932, + "learning_rate": 0.00041899261080029904, + "loss": 3.5839173793792725, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.020935002714395523, + "learning_rate": 0.00041874714909679754, + "loss": 3.5794217586517334, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.020799759775400162, + "learning_rate": 0.00041850159413790863, + "loss": 3.6033859252929688, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.019480884075164795, + "learning_rate": 0.0004182559461214715, + "loss": 3.580097198486328, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.019300777465105057, + "learning_rate": 0.0004180102052454004, + "loss": 3.550474166870117, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.020265772938728333, + "learning_rate": 0.00041776437170768427, + "loss": 3.5711655616760254, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.019972780719399452, + "learning_rate": 0.0004175184457063868, + "loss": 3.5636885166168213, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.019635623320937157, + "learning_rate": 0.000417272427439646, + "loss": 3.5851311683654785, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.021451901644468307, + "learning_rate": 0.00041702631710567443, + "loss": 3.5956759452819824, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.019456414505839348, + "learning_rate": 0.00041678011490275875, + "loss": 3.595731258392334, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.019757689908146858, + "learning_rate": 0.00041653382102925957, + "loss": 3.569652557373047, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.020571604371070862, + "learning_rate": 0.00041628743568361147, + "loss": 3.6007747650146484, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.02153848111629486, + "learning_rate": 0.00041604095906432265, + "loss": 3.5915260314941406, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.020875850692391396, + "learning_rate": 0.00041579439136997493, + "loss": 3.5582690238952637, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.02089419960975647, + "learning_rate": 0.00041554773279922337, + "loss": 3.5756618976593018, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01961098238825798, + "learning_rate": 0.0004153009835507963, + "loss": 3.5918776988983154, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.021075716242194176, + "learning_rate": 0.0004150541438234952, + "loss": 3.5342202186584473, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.020745476707816124, + "learning_rate": 0.00041480721381619435, + "loss": 3.5976028442382812, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.02001352794468403, + "learning_rate": 0.00041456019372784086, + "loss": 3.5686182975769043, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.021684160456061363, + "learning_rate": 0.0004143130837574543, + "loss": 3.553445339202881, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.022551920264959335, + "learning_rate": 0.0004140658841041267, + "loss": 3.5712709426879883, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.022758224979043007, + "learning_rate": 0.0004138185949670224, + "loss": 3.6101667881011963, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.020714636892080307, + "learning_rate": 0.00041357121654537766, + "loss": 3.5740506649017334, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.022694187238812447, + "learning_rate": 0.00041332374903850104, + "loss": 3.5663979053497314, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.021664340049028397, + "learning_rate": 0.00041307619264577234, + "loss": 3.5830888748168945, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.019072270020842552, + "learning_rate": 0.0004128285475666436, + "loss": 3.5861973762512207, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.020153764635324478, + "learning_rate": 0.0004125808140006378, + "loss": 3.566444158554077, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019666558131575584, + "learning_rate": 0.0004123329921473494, + "loss": 3.5789592266082764, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.02101624570786953, + "learning_rate": 0.0004120850822064439, + "loss": 3.551722526550293, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.01951020397245884, + "learning_rate": 0.00041183708437765815, + "loss": 3.6069014072418213, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.01985476352274418, + "learning_rate": 0.00041158899886079925, + "loss": 3.5944924354553223, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02195335365831852, + "learning_rate": 0.00041134082585574534, + "loss": 3.5828046798706055, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02105172537267208, + "learning_rate": 0.00041109256556244497, + "loss": 3.5830953121185303, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.01888827420771122, + "learning_rate": 0.00041084421818091685, + "loss": 3.5656847953796387, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.017867228016257286, + "learning_rate": 0.0004105957839112501, + "loss": 3.5792243480682373, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.018577905371785164, + "learning_rate": 0.0004103472629536036, + "loss": 3.56992506980896, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.020045632496476173, + "learning_rate": 0.0004100986555082063, + "loss": 3.590876579284668, + "step": 1440 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6.139226722664448e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-1440/training_args.bin b/runs/l2r90-baseline/checkpoint-1440/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1440/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-1800/chat_template.jinja b/runs/l2r90-baseline/checkpoint-1800/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-1800/config.json b/runs/l2r90-baseline/checkpoint-1800/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-1800/generation_config.json b/runs/l2r90-baseline/checkpoint-1800/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-1800/model.safetensors b/runs/l2r90-baseline/checkpoint-1800/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c187dd9d2d4e7a90cbe7b8e403678f7314cb6a0e --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18884bb3a15b51ad77d907f70762bbd36667561ebe02aca38a16037270e9ec38 +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-1800/optimizer.pt b/runs/l2r90-baseline/checkpoint-1800/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..46c0412ddcfa76dcac5435544f340cfe4b3ef7fb --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3194f6e151dbbffe602de2085a569217940541b1fea695aca41a784822b50e34 +size 1166825338 diff --git a/runs/l2r90-baseline/checkpoint-1800/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-1800/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..0e1e82cfece0f9f4b960303264240de21a41f2b2 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8303cc5f635d6851ead500cf1dcb6114b8355d49a2ca28015d085bed26a36524 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1800/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-1800/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..4b06db675e31ec9653c89af439b51ab40f3cdab3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99e1b9386d1bf84a8e2951517d247f0644fbc1877429b25df213b0870f58484c +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1800/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-1800/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..88526073cc9c988c98d9074803bd3fbc38252bdd --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb42dab9f2cc8d80099e5d46382913710a69bfab9566367bb8641baf78018e00 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1800/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-1800/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..a9ec1a11b85512a95c1b62566c99d6bf9aa9c3ee --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34a8598ceeb75c2fc78a1e5b43d2d5910250c2ea2784dd658d38680cdcec97b0 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-1800/scheduler.pt b/runs/l2r90-baseline/checkpoint-1800/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..6e0ba2afe167e5831737cd4b431b166033f17fab --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee9cab552a4cabfd465ec6f079c65fbe7304d4a3df75e2e2b3e2eb614e447c70 +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-1800/tokenizer.json b/runs/l2r90-baseline/checkpoint-1800/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-1800/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-1800/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-1800/trainer_state.json b/runs/l2r90-baseline/checkpoint-1800/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..686bf8e16127bbbdb9711812f708be77cdd88118 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/trainer_state.json @@ -0,0 +1,12826 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 25.0, + "eval_steps": 500, + "global_step": 1800, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 5.549652099609375, + "eval_runtime": 57.761, + "eval_samples_per_second": 42.278, + "eval_steps_per_second": 1.333, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.03958690166473389, + "learning_rate": 0.0005919487226162231, + "loss": 5.459286689758301, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.03299890458583832, + "learning_rate": 0.0005918869546699877, + "loss": 5.45464563369751, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.028040863573551178, + "learning_rate": 0.0005918249539727509, + "loss": 5.444796562194824, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.028365720063447952, + "learning_rate": 0.0005917627205744655, + "loss": 5.447852611541748, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.02822224237024784, + "learning_rate": 0.0005917002545252719, + "loss": 5.453883171081543, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.036075398325920105, + "learning_rate": 0.0005916375558754977, + "loss": 5.462671279907227, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04683876410126686, + "learning_rate": 0.0005915746246756581, + "loss": 5.443917274475098, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.05457588657736778, + "learning_rate": 0.0005915114609764558, + "loss": 5.444374084472656, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.04250836744904518, + "learning_rate": 0.0005914480648287804, + "loss": 5.435029983520508, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.029477844014763832, + "learning_rate": 0.0005913844362837093, + "loss": 5.421862602233887, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.03673284500837326, + "learning_rate": 0.0005913205753925066, + "loss": 5.430391311645508, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.030489861965179443, + "learning_rate": 0.0005912564822066241, + "loss": 5.389314651489258, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.04037046805024147, + "learning_rate": 0.0005911921567777005, + "loss": 5.425060272216797, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.0445217490196228, + "learning_rate": 0.0005911275991575613, + "loss": 5.437565326690674, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.03885014355182648, + "learning_rate": 0.0005910628093982198, + "loss": 5.410930633544922, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.0334964245557785, + "learning_rate": 0.0005909977875518759, + "loss": 5.412092208862305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.03499286621809006, + "learning_rate": 0.0005909325336709164, + "loss": 5.382216453552246, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.04171644523739815, + "learning_rate": 0.0005908670478079152, + "loss": 5.387385368347168, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.057558078318834305, + "learning_rate": 0.000590801330015633, + "loss": 5.437925815582275, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.053846780210733414, + "learning_rate": 0.0005907353803470177, + "loss": 5.41644811630249, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.027259036898612976, + "learning_rate": 0.0005906691988552034, + "loss": 5.409185409545898, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.03394043818116188, + "learning_rate": 0.0005906027855935115, + "loss": 5.377973556518555, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.03731574863195419, + "learning_rate": 0.0005905361406154501, + "loss": 5.382505893707275, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.04705141484737396, + "learning_rate": 0.0005904692639747137, + "loss": 5.413661003112793, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.044811323285102844, + "learning_rate": 0.0005904021557251837, + "loss": 5.417835235595703, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.04000752419233322, + "learning_rate": 0.0005903348159209277, + "loss": 5.406160354614258, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.038364164531230927, + "learning_rate": 0.0005902672446162007, + "loss": 5.357950687408447, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.030500127002596855, + "learning_rate": 0.0005901994418654432, + "loss": 5.344951629638672, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.030644729733467102, + "learning_rate": 0.0005901314077232829, + "loss": 5.3599348068237305, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.037976790219545364, + "learning_rate": 0.0005900631422445335, + "loss": 5.354084491729736, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.04106011986732483, + "learning_rate": 0.0005899946454841955, + "loss": 5.363678455352783, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.04334832355380058, + "learning_rate": 0.0005899259174974553, + "loss": 5.350959300994873, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.049939028918743134, + "learning_rate": 0.000589856958339686, + "loss": 5.3317155838012695, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.05300058051943779, + "learning_rate": 0.0005897877680664465, + "loss": 5.353314399719238, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.05943184345960617, + "learning_rate": 0.0005897183467334822, + "loss": 5.3683929443359375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.05600789934396744, + "learning_rate": 0.0005896486943967249, + "loss": 5.345516204833984, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.05392718315124512, + "learning_rate": 0.0005895788111122919, + "loss": 5.291389465332031, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.037913475185632706, + "learning_rate": 0.0005895086969364871, + "loss": 5.353423118591309, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.03362123295664787, + "learning_rate": 0.0005894383519258001, + "loss": 5.379500389099121, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032558735460042953, + "learning_rate": 0.0005893677761369066, + "loss": 5.34043025970459, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.04074402526021004, + "learning_rate": 0.0005892969696266683, + "loss": 5.303869247436523, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.04270484298467636, + "learning_rate": 0.0005892259324521328, + "loss": 5.27297830581665, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.04604959115386009, + "learning_rate": 0.0005891546646705334, + "loss": 5.313268184661865, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.04462246224284172, + "learning_rate": 0.0005890831663392891, + "loss": 5.31167459487915, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.04966094344854355, + "learning_rate": 0.0005890114375160052, + "loss": 5.316695690155029, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.04662073776125908, + "learning_rate": 0.0005889394782584718, + "loss": 5.316898822784424, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.04380315542221069, + "learning_rate": 0.0005888672886246656, + "loss": 5.285087585449219, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03239579498767853, + "learning_rate": 0.0005887948686727483, + "loss": 5.304283142089844, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.029690183699131012, + "learning_rate": 0.0005887222184610675, + "loss": 5.305715084075928, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.03444967418909073, + "learning_rate": 0.0005886493380481559, + "loss": 5.2707672119140625, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.03302565962076187, + "learning_rate": 0.0005885762274927322, + "loss": 5.277251243591309, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.028477763757109642, + "learning_rate": 0.0005885028868537, + "loss": 5.295318603515625, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.022988801822066307, + "learning_rate": 0.0005884293161901487, + "loss": 5.249057292938232, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.026115216314792633, + "learning_rate": 0.000588355515561353, + "loss": 5.270636081695557, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.02576332725584507, + "learning_rate": 0.0005882814850267723, + "loss": 5.263096332550049, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.023978352546691895, + "learning_rate": 0.0005882072246460521, + "loss": 5.276605606079102, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.020695513114333153, + "learning_rate": 0.0005881327344790223, + "loss": 5.2141432762146, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01952219195663929, + "learning_rate": 0.0005880580145856986, + "loss": 5.214289665222168, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.022633206099271774, + "learning_rate": 0.0005879830650262813, + "loss": 5.236690998077393, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.023082347586750984, + "learning_rate": 0.0005879078858611557, + "loss": 5.2092156410217285, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.019620463252067566, + "learning_rate": 0.0005878324771508927, + "loss": 5.264451503753662, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.021504493430256844, + "learning_rate": 0.0005877568389562475, + "loss": 5.23444938659668, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.025672364979982376, + "learning_rate": 0.0005876809713381606, + "loss": 5.205289363861084, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.034436631947755814, + "learning_rate": 0.0005876048743577569, + "loss": 5.2047648429870605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.042389530688524246, + "learning_rate": 0.0005875285480763466, + "loss": 5.192008018493652, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.03310716897249222, + "learning_rate": 0.0005874519925554244, + "loss": 5.162734031677246, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.03029288910329342, + "learning_rate": 0.0005873752078566694, + "loss": 5.212151527404785, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04619310051202774, + "learning_rate": 0.0005872981940419459, + "loss": 5.215909004211426, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.05618087574839592, + "learning_rate": 0.0005872209511733027, + "loss": 5.210192680358887, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.03783707693219185, + "learning_rate": 0.0005871434793129726, + "loss": 5.201597690582275, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.03721104562282562, + "learning_rate": 0.0005870657785233735, + "loss": 5.188891410827637, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.033336155116558075, + "learning_rate": 0.0005869878488671075, + "loss": 5.1598615646362305, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.2592363357543945, + "eval_runtime": 57.3343, + "eval_samples_per_second": 42.592, + "eval_steps_per_second": 1.343, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03719741478562355, + "learning_rate": 0.0005869096904069611, + "loss": 5.1646623611450195, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.04888090118765831, + "learning_rate": 0.0005868313032059052, + "loss": 5.171182155609131, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.05073149502277374, + "learning_rate": 0.0005867526873270949, + "loss": 5.183949947357178, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.04187731072306633, + "learning_rate": 0.0005866738428338695, + "loss": 5.176133155822754, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.039445552974939346, + "learning_rate": 0.000586594769789753, + "loss": 5.148100852966309, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.035551708191633224, + "learning_rate": 0.0005865154682584524, + "loss": 5.151512145996094, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.031175289303064346, + "learning_rate": 0.0005864359383038602, + "loss": 5.102468967437744, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.03696789592504501, + "learning_rate": 0.000586356179990052, + "loss": 5.119902610778809, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.04836135357618332, + "learning_rate": 0.0005862761933812875, + "loss": 5.15183162689209, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.05206914246082306, + "learning_rate": 0.0005861959785420106, + "loss": 5.154870986938477, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.046368084847927094, + "learning_rate": 0.000586115535536849, + "loss": 5.130651473999023, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.07146432995796204, + "learning_rate": 0.000586034864430614, + "loss": 5.175804138183594, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.08809789270162582, + "learning_rate": 0.0005859539652883008, + "loss": 5.229716777801514, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.08116360008716583, + "learning_rate": 0.0005858728381750888, + "loss": 5.21910285949707, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.08282570540904999, + "learning_rate": 0.0005857914831563402, + "loss": 5.2336506843566895, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.07403547316789627, + "learning_rate": 0.0005857099002976016, + "loss": 5.251604080200195, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.07024004310369492, + "learning_rate": 0.0005856280896646026, + "loss": 5.267114162445068, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.056330062448978424, + "learning_rate": 0.0005855460513232567, + "loss": 5.201976299285889, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.051763955503702164, + "learning_rate": 0.0005854637853396606, + "loss": 5.1733198165893555, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.039900992065668106, + "learning_rate": 0.0005853812917800945, + "loss": 5.222941875457764, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.04187863692641258, + "learning_rate": 0.0005852985707110221, + "loss": 5.165033340454102, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.042645812034606934, + "learning_rate": 0.0005852156221990901, + "loss": 5.157554626464844, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.03856075555086136, + "learning_rate": 0.0005851324463111289, + "loss": 5.184629440307617, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.029387330636382103, + "learning_rate": 0.0005850490431141516, + "loss": 5.118739128112793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.027584781870245934, + "learning_rate": 0.0005849654126753545, + "loss": 5.100855350494385, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.023655924946069717, + "learning_rate": 0.0005848815550621175, + "loss": 5.154520034790039, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.024163050577044487, + "learning_rate": 0.0005847974703420028, + "loss": 5.105189323425293, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.020369920879602432, + "learning_rate": 0.0005847131585827564, + "loss": 5.113671779632568, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.021030737087130547, + "learning_rate": 0.0005846286198523061, + "loss": 5.0847649574279785, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.01729249767959118, + "learning_rate": 0.0005845438542187638, + "loss": 5.053225517272949, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.018704712390899658, + "learning_rate": 0.0005844588617504236, + "loss": 5.079106330871582, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.02049064077436924, + "learning_rate": 0.0005843736425157621, + "loss": 5.044185161590576, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.020736856386065483, + "learning_rate": 0.000584288196583439, + "loss": 5.041932106018066, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.016543276607990265, + "learning_rate": 0.0005842025240222966, + "loss": 5.032522201538086, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.018414106220006943, + "learning_rate": 0.0005841166249013598, + "loss": 5.037397384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.01611698977649212, + "learning_rate": 0.0005840304992898359, + "loss": 5.038179397583008, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.014310561120510101, + "learning_rate": 0.0005839441472571147, + "loss": 5.0146484375, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.015472214668989182, + "learning_rate": 0.0005838575688727685, + "loss": 5.060323715209961, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.016373401507735252, + "learning_rate": 0.0005837707642065518, + "loss": 5.032100677490234, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.018694665282964706, + "learning_rate": 0.000583683733328402, + "loss": 5.036993026733398, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.019977087154984474, + "learning_rate": 0.0005835964763084378, + "loss": 4.991037368774414, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.023741086944937706, + "learning_rate": 0.000583508993216961, + "loss": 5.043832778930664, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.025061562657356262, + "learning_rate": 0.0005834212841244548, + "loss": 5.016516208648682, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.020848384127020836, + "learning_rate": 0.0005833333491015851, + "loss": 4.986289978027344, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.015086354687809944, + "learning_rate": 0.0005832451882191995, + "loss": 4.9948625564575195, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.015096964314579964, + "learning_rate": 0.0005831568015483274, + "loss": 4.996342658996582, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016932200640439987, + "learning_rate": 0.0005830681891601807, + "loss": 4.998758316040039, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.019936930388212204, + "learning_rate": 0.0005829793511261525, + "loss": 4.959854602813721, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.027658682316541672, + "learning_rate": 0.0005828902875178183, + "loss": 4.968629837036133, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.03299275413155556, + "learning_rate": 0.0005828009984069347, + "loss": 4.994814872741699, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.033717650920152664, + "learning_rate": 0.0005827114838654406, + "loss": 4.976295471191406, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.02969924360513687, + "learning_rate": 0.0005826217439654562, + "loss": 4.969155788421631, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.02846827358007431, + "learning_rate": 0.0005825317787792831, + "loss": 4.989721298217773, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.03670600801706314, + "learning_rate": 0.0005824415883794049, + "loss": 4.988689422607422, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.045570675283670425, + "learning_rate": 0.0005823511728384863, + "loss": 4.974407196044922, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.04564111679792404, + "learning_rate": 0.0005822605322293733, + "loss": 4.966331481933594, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.04173210263252258, + "learning_rate": 0.0005821696666250937, + "loss": 4.925197601318359, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.04228242486715317, + "learning_rate": 0.0005820785760988559, + "loss": 4.965020656585693, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.05144782364368439, + "learning_rate": 0.0005819872607240503, + "loss": 5.0022149085998535, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.049400657415390015, + "learning_rate": 0.0005818957205742478, + "loss": 4.9635114669799805, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.043692708015441895, + "learning_rate": 0.0005818039557232005, + "loss": 4.977723121643066, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.04891742020845413, + "learning_rate": 0.0005817119662448421, + "loss": 4.996312141418457, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.035564981400966644, + "learning_rate": 0.0005816197522132866, + "loss": 4.940932273864746, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04679827764630318, + "learning_rate": 0.0005815273137028292, + "loss": 4.983131408691406, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.05489890277385712, + "learning_rate": 0.0005814346507879459, + "loss": 4.977827548980713, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04640809819102287, + "learning_rate": 0.0005813417635432937, + "loss": 4.954761505126953, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04439452290534973, + "learning_rate": 0.00058124865204371, + "loss": 4.914865970611572, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.04587096720933914, + "learning_rate": 0.0005811553163642131, + "loss": 4.949592590332031, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.044119734317064285, + "learning_rate": 0.0005810617565800019, + "loss": 4.956068992614746, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.043133001774549484, + "learning_rate": 0.0005809679727664559, + "loss": 4.993522644042969, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04315278306603432, + "learning_rate": 0.0005808739649991348, + "loss": 4.894291877746582, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04819472134113312, + "learning_rate": 0.0005807797333537792, + "loss": 4.958306789398193, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.023706436157227, + "eval_runtime": 57.4082, + "eval_samples_per_second": 42.537, + "eval_steps_per_second": 1.341, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.04403403028845787, + "learning_rate": 0.0005806852779063098, + "loss": 4.884971618652344, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.041993748396635056, + "learning_rate": 0.0005805905987328275, + "loss": 4.916353225708008, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.042302004992961884, + "learning_rate": 0.0005804956959096138, + "loss": 4.9012556076049805, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.044500596821308136, + "learning_rate": 0.0005804005695131299, + "loss": 4.915240287780762, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.03914736211299896, + "learning_rate": 0.0005803052196200177, + "loss": 4.885915756225586, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.0361703559756279, + "learning_rate": 0.0005802096463070988, + "loss": 4.829287528991699, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03432302549481392, + "learning_rate": 0.0005801138496513749, + "loss": 4.926834583282471, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.0317804180085659, + "learning_rate": 0.0005800178297300276, + "loss": 4.8707170486450195, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.035261351615190506, + "learning_rate": 0.0005799215866204183, + "loss": 4.897791862487793, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.03276067599654198, + "learning_rate": 0.0005798251204000886, + "loss": 4.864504814147949, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.034757066518068314, + "learning_rate": 0.0005797284311467594, + "loss": 4.890812873840332, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.033726081252098083, + "learning_rate": 0.0005796315189383316, + "loss": 4.831684112548828, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.03074726276099682, + "learning_rate": 0.0005795343838528855, + "loss": 4.8534111976623535, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02215893566608429, + "learning_rate": 0.0005794370259686811, + "loss": 4.8311238288879395, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.02132033370435238, + "learning_rate": 0.0005793394453641581, + "loss": 4.842617034912109, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020825816318392754, + "learning_rate": 0.0005792416421179351, + "loss": 4.84535026550293, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.019334357231855392, + "learning_rate": 0.0005791436163088106, + "loss": 4.839937686920166, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017372334375977516, + "learning_rate": 0.0005790453680157622, + "loss": 4.8115553855896, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.0190406683832407, + "learning_rate": 0.0005789468973179468, + "loss": 4.832542419433594, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.017372574657201767, + "learning_rate": 0.0005788482042947004, + "loss": 4.793632507324219, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.019171632826328278, + "learning_rate": 0.0005787492890255382, + "loss": 4.776987552642822, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.02109278179705143, + "learning_rate": 0.0005786501515901545, + "loss": 4.8216657638549805, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.026234494522213936, + "learning_rate": 0.0005785507920684225, + "loss": 4.8109049797058105, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.0357641875743866, + "learning_rate": 0.0005784512105403944, + "loss": 4.821109771728516, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.03627947345376015, + "learning_rate": 0.000578351407086301, + "loss": 4.810632228851318, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.032253146171569824, + "learning_rate": 0.0005782513817865527, + "loss": 4.7930521965026855, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.03402210772037506, + "learning_rate": 0.0005781511347217375, + "loss": 4.772500991821289, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.0321584977209568, + "learning_rate": 0.000578050665972623, + "loss": 4.781399726867676, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.030242707580327988, + "learning_rate": 0.0005779499756201549, + "loss": 4.802542686462402, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.02361336164176464, + "learning_rate": 0.0005778490637454576, + "loss": 4.771912574768066, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.028781302273273468, + "learning_rate": 0.000577747930429834, + "loss": 4.796924591064453, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.027042683213949203, + "learning_rate": 0.0005776465757547654, + "loss": 4.745218276977539, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.02715984545648098, + "learning_rate": 0.0005775449998019113, + "loss": 4.792944431304932, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.0314096100628376, + "learning_rate": 0.00057744320265311, + "loss": 4.7649993896484375, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.028749193996191025, + "learning_rate": 0.0005773411843903773, + "loss": 4.7980546951293945, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.02655309997498989, + "learning_rate": 0.0005772389450959075, + "loss": 4.746520519256592, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.027497397735714912, + "learning_rate": 0.000577136484852073, + "loss": 4.7490715980529785, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.0343393050134182, + "learning_rate": 0.000577033803741424, + "loss": 4.74436092376709, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.04444412514567375, + "learning_rate": 0.0005769309018466891, + "loss": 4.755973815917969, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.0437556728720665, + "learning_rate": 0.0005768277792507744, + "loss": 4.784368515014648, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0369921550154686, + "learning_rate": 0.0005767244360367638, + "loss": 4.753410339355469, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.05038141459226608, + "learning_rate": 0.0005766208722879192, + "loss": 4.755491256713867, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.04511422663927078, + "learning_rate": 0.00057651708808768, + "loss": 4.760293006896973, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.048762883991003036, + "learning_rate": 0.0005764130835196632, + "loss": 4.755958080291748, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.06749686598777771, + "learning_rate": 0.0005763088586676634, + "loss": 4.802557468414307, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.06451250612735748, + "learning_rate": 0.0005762044136156528, + "loss": 4.812925338745117, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.05436156690120697, + "learning_rate": 0.0005760997484477809, + "loss": 4.807103157043457, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.05233754590153694, + "learning_rate": 0.0005759948632483742, + "loss": 4.782449722290039, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.05519746616482735, + "learning_rate": 0.0005758897581019371, + "loss": 4.808449745178223, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.0578567236661911, + "learning_rate": 0.000575784433093151, + "loss": 4.7676005363464355, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.0588264986872673, + "learning_rate": 0.0005756788883068743, + "loss": 4.805377960205078, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04225790873169899, + "learning_rate": 0.0005755731238281423, + "loss": 4.748775005340576, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.0376368910074234, + "learning_rate": 0.0005754671397421678, + "loss": 4.759791374206543, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03830771893262863, + "learning_rate": 0.0005753609361343402, + "loss": 4.729974746704102, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.03684520348906517, + "learning_rate": 0.0005752545130902256, + "loss": 4.729854583740234, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02983408235013485, + "learning_rate": 0.0005751478706955674, + "loss": 4.78502082824707, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02854904904961586, + "learning_rate": 0.0005750410090362854, + "loss": 4.717472076416016, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.023819150403141975, + "learning_rate": 0.0005749339281984761, + "loss": 4.717124938964844, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.02477158047258854, + "learning_rate": 0.0005748266282684124, + "loss": 4.71781063079834, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.022002534940838814, + "learning_rate": 0.0005747191093325443, + "loss": 4.7480669021606445, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.02133483812212944, + "learning_rate": 0.0005746113714774976, + "loss": 4.693770408630371, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.022407138720154762, + "learning_rate": 0.0005745034147900747, + "loss": 4.720734119415283, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.020850926637649536, + "learning_rate": 0.0005743952393572544, + "loss": 4.745505332946777, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.020420582965016365, + "learning_rate": 0.0005742868452661918, + "loss": 4.688610076904297, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.019014107063412666, + "learning_rate": 0.0005741782326042181, + "loss": 4.673436164855957, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.017393039539456367, + "learning_rate": 0.0005740694014588403, + "loss": 4.6941118240356445, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.017877329140901566, + "learning_rate": 0.0005739603519177419, + "loss": 4.667474269866943, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.017749635502696037, + "learning_rate": 0.0005738510840687821, + "loss": 4.6901092529296875, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.018746376037597656, + "learning_rate": 0.000573741597999996, + "loss": 4.680166721343994, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016260217875242233, + "learning_rate": 0.0005736318937995947, + "loss": 4.6799821853637695, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015099814161658287, + "learning_rate": 0.0005735219715559646, + "loss": 4.680666923522949, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.016539735719561577, + "learning_rate": 0.0005734118313576683, + "loss": 4.639509201049805, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 4.768754482269287, + "eval_runtime": 58.3352, + "eval_samples_per_second": 41.862, + "eval_steps_per_second": 1.32, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.019733518362045288, + "learning_rate": 0.0005733014732934436, + "loss": 4.635312080383301, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.022571036592125893, + "learning_rate": 0.0005731908974522042, + "loss": 4.6041765213012695, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.022238552570343018, + "learning_rate": 0.0005730801039230389, + "loss": 4.648976802825928, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.02115180902183056, + "learning_rate": 0.0005729690927952123, + "loss": 4.623968124389648, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.021789928898215294, + "learning_rate": 0.0005728578641581637, + "loss": 4.620028018951416, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.02593887411057949, + "learning_rate": 0.0005727464181015081, + "loss": 4.591062545776367, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.0347842313349247, + "learning_rate": 0.0005726347547150357, + "loss": 4.632149696350098, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04119519516825676, + "learning_rate": 0.0005725228740887117, + "loss": 4.6254191398620605, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.04895750433206558, + "learning_rate": 0.0005724107763126761, + "loss": 4.621372222900391, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.05752161890268326, + "learning_rate": 0.0005722984614772442, + "loss": 4.61965274810791, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05115870013833046, + "learning_rate": 0.000572185929672906, + "loss": 4.63119649887085, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.05721089243888855, + "learning_rate": 0.0005720731809903263, + "loss": 4.652647018432617, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.0646839588880539, + "learning_rate": 0.000571960215520345, + "loss": 4.642184257507324, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.0682854950428009, + "learning_rate": 0.0005718470333539757, + "loss": 4.676589012145996, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.057402849197387695, + "learning_rate": 0.0005717336345824077, + "loss": 4.6585283279418945, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.05885373800992966, + "learning_rate": 0.0005716200192970043, + "loss": 4.671728134155273, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.05265705659985542, + "learning_rate": 0.0005715061875893032, + "loss": 4.641249656677246, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.05417842045426369, + "learning_rate": 0.0005713921395510166, + "loss": 4.661381721496582, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.06602369248867035, + "learning_rate": 0.0005712778752740307, + "loss": 4.684137344360352, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.06004182621836662, + "learning_rate": 0.0005711633948504066, + "loss": 4.625332832336426, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.04801511764526367, + "learning_rate": 0.0005710486983723787, + "loss": 4.664791107177734, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.04380105435848236, + "learning_rate": 0.0005709337859323561, + "loss": 4.664966583251953, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.03858727216720581, + "learning_rate": 0.0005708186576229218, + "loss": 4.661867618560791, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03240576758980751, + "learning_rate": 0.0005707033135368323, + "loss": 4.5855326652526855, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.028561925515532494, + "learning_rate": 0.0005705877537670184, + "loss": 4.637331008911133, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.026389246806502342, + "learning_rate": 0.0005704719784065846, + "loss": 4.598666667938232, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.02484819106757641, + "learning_rate": 0.0005703559875488088, + "loss": 4.601729393005371, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.02522977627813816, + "learning_rate": 0.0005702397812871429, + "loss": 4.609833717346191, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02589508146047592, + "learning_rate": 0.0005701233597152121, + "loss": 4.574592590332031, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.023374728858470917, + "learning_rate": 0.0005700067229268154, + "loss": 4.61362361907959, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.02465851791203022, + "learning_rate": 0.0005698898710159245, + "loss": 4.639899253845215, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.02258715033531189, + "learning_rate": 0.0005697728040766852, + "loss": 4.5656538009643555, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.018009789288043976, + "learning_rate": 0.0005696555222034162, + "loss": 4.61449670791626, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02121281623840332, + "learning_rate": 0.0005695380254906094, + "loss": 4.553830623626709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.018078701570630074, + "learning_rate": 0.0005694203140329296, + "loss": 4.534579277038574, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.0169303547590971, + "learning_rate": 0.000569302387925215, + "loss": 4.533995628356934, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.016318142414093018, + "learning_rate": 0.0005691842472624764, + "loss": 4.544438362121582, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.016664575785398483, + "learning_rate": 0.0005690658921398977, + "loss": 4.5913286209106445, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.01641504280269146, + "learning_rate": 0.0005689473226528354, + "loss": 4.520743370056152, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.017012735828757286, + "learning_rate": 0.0005688285388968187, + "loss": 4.517453193664551, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01672312617301941, + "learning_rate": 0.0005687095409675499, + "loss": 4.5235915184021, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.016893865540623665, + "learning_rate": 0.000568590328960903, + "loss": 4.543200492858887, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.017530135810375214, + "learning_rate": 0.0005684709029729253, + "loss": 4.54912805557251, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.018698599189519882, + "learning_rate": 0.0005683512630998361, + "loss": 4.522650718688965, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01984071172773838, + "learning_rate": 0.000568231409438027, + "loss": 4.546173095703125, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.023218289017677307, + "learning_rate": 0.0005681113420840619, + "loss": 4.5305094718933105, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.028968270868062973, + "learning_rate": 0.000567991061134677, + "loss": 4.521440505981445, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.032370273023843765, + "learning_rate": 0.0005678705666867805, + "loss": 4.54168701171875, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.033967748284339905, + "learning_rate": 0.0005677498588374524, + "loss": 4.538717269897461, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.028757929801940918, + "learning_rate": 0.0005676289376839452, + "loss": 4.517228603363037, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.029757825657725334, + "learning_rate": 0.0005675078033236827, + "loss": 4.510310173034668, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.03581003099679947, + "learning_rate": 0.0005673864558542605, + "loss": 4.506901264190674, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.03483985736966133, + "learning_rate": 0.0005672648953734462, + "loss": 4.549284934997559, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.031016027554869652, + "learning_rate": 0.000567143121979179, + "loss": 4.530084609985352, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.029373526573181152, + "learning_rate": 0.0005670211357695693, + "loss": 4.492218971252441, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.031661372631788254, + "learning_rate": 0.0005668989368428994, + "loss": 4.502824783325195, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.027251387014985085, + "learning_rate": 0.0005667765252976227, + "loss": 4.49193811416626, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02263176068663597, + "learning_rate": 0.0005666539012323639, + "loss": 4.512019157409668, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.02355228364467621, + "learning_rate": 0.0005665310647459189, + "loss": 4.508917331695557, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.023297371342778206, + "learning_rate": 0.0005664080159372551, + "loss": 4.497771263122559, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.019364114850759506, + "learning_rate": 0.0005662847549055107, + "loss": 4.51470947265625, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.023899901658296585, + "learning_rate": 0.0005661612817499947, + "loss": 4.48771858215332, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.02330983616411686, + "learning_rate": 0.0005660375965701872, + "loss": 4.478496551513672, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.027252234518527985, + "learning_rate": 0.0005659136994657392, + "loss": 4.497672080993652, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.030886363238096237, + "learning_rate": 0.0005657895905364723, + "loss": 4.49789571762085, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02515559457242489, + "learning_rate": 0.0005656652698823788, + "loss": 4.490771293640137, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020107125863432884, + "learning_rate": 0.0005655407376036217, + "loss": 4.481879234313965, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.022076716646552086, + "learning_rate": 0.0005654159938005345, + "loss": 4.494697570800781, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.025525815784931183, + "learning_rate": 0.0005652910385736207, + "loss": 4.517453193664551, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.03005894646048546, + "learning_rate": 0.0005651658720235548, + "loss": 4.453815460205078, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.030178502202033997, + "learning_rate": 0.0005650404942511812, + "loss": 4.493922233581543, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.029249394312500954, + "learning_rate": 0.0005649149053575141, + "loss": 4.453763008117676, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 4.607339382171631, + "eval_runtime": 57.9216, + "eval_samples_per_second": 42.16, + "eval_steps_per_second": 1.329, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.03362950682640076, + "learning_rate": 0.0005647891054437386, + "loss": 4.430700302124023, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.04223684221506119, + "learning_rate": 0.0005646630946112096, + "loss": 4.486915588378906, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.05192171037197113, + "learning_rate": 0.0005645368729614512, + "loss": 4.467541694641113, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.042525868862867355, + "learning_rate": 0.0005644104405961582, + "loss": 4.45252799987793, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.0406370609998703, + "learning_rate": 0.0005642837976171949, + "loss": 4.414548873901367, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.044476233422756195, + "learning_rate": 0.0005641569441265952, + "loss": 4.453492164611816, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.04687326401472092, + "learning_rate": 0.0005640298802265626, + "loss": 4.477725028991699, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.04011611267924309, + "learning_rate": 0.0005639026060194704, + "loss": 4.423356533050537, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.04971884936094284, + "learning_rate": 0.0005637751216078609, + "loss": 4.4853515625, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.05863234028220177, + "learning_rate": 0.0005636474270944461, + "loss": 4.466001033782959, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.057710036635398865, + "learning_rate": 0.0005635195225821072, + "loss": 4.466723442077637, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.05155074968934059, + "learning_rate": 0.0005633914081738945, + "loss": 4.4823198318481445, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.04535634070634842, + "learning_rate": 0.0005632630839730275, + "loss": 4.471219062805176, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.045147854834795, + "learning_rate": 0.0005631345500828946, + "loss": 4.476888179779053, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.03730619326233864, + "learning_rate": 0.0005630058066070533, + "loss": 4.428932189941406, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.034599728882312775, + "learning_rate": 0.0005628768536492299, + "loss": 4.47627592086792, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.032904356718063354, + "learning_rate": 0.0005627476913133193, + "loss": 4.422671794891357, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.03356337174773216, + "learning_rate": 0.0005626183197033856, + "loss": 4.456406593322754, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.035042595118284225, + "learning_rate": 0.0005624887389236609, + "loss": 4.371820449829102, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.03269032761454582, + "learning_rate": 0.0005623589490785462, + "loss": 4.402605056762695, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.028261031955480576, + "learning_rate": 0.0005622289502726108, + "loss": 4.433136940002441, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03301700949668884, + "learning_rate": 0.0005620987426105925, + "loss": 4.388359546661377, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.031664177775382996, + "learning_rate": 0.000561968326197397, + "loss": 4.41859245300293, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.026281699538230896, + "learning_rate": 0.0005618377011380988, + "loss": 4.406128883361816, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.023097911849617958, + "learning_rate": 0.0005617068675379398, + "loss": 4.389717102050781, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.022349286824464798, + "learning_rate": 0.0005615758255023306, + "loss": 4.408173561096191, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.021958226338028908, + "learning_rate": 0.0005614445751368491, + "loss": 4.434830188751221, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024551713839173317, + "learning_rate": 0.0005613131165472415, + "loss": 4.386944770812988, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.028048967942595482, + "learning_rate": 0.0005611814498394216, + "loss": 4.402279376983643, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02733907848596573, + "learning_rate": 0.0005610495751194708, + "loss": 4.403069496154785, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.020858559757471085, + "learning_rate": 0.0005609174924936384, + "loss": 4.406920909881592, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.021491149440407753, + "learning_rate": 0.0005607852020683409, + "loss": 4.385799407958984, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.021495329216122627, + "learning_rate": 0.0005606527039501621, + "loss": 4.4252495765686035, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.019708355888724327, + "learning_rate": 0.0005605199982458535, + "loss": 4.388864994049072, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.01886274851858616, + "learning_rate": 0.0005603870850623338, + "loss": 4.335962295532227, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.01955341547727585, + "learning_rate": 0.0005602539645066884, + "loss": 4.373560905456543, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.018962012603878975, + "learning_rate": 0.0005601206366861706, + "loss": 4.380022048950195, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.016663335263729095, + "learning_rate": 0.0005599871017081999, + "loss": 4.421034336090088, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01676873490214348, + "learning_rate": 0.0005598533596803631, + "loss": 4.382775783538818, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.014379126951098442, + "learning_rate": 0.0005597194107104137, + "loss": 4.41373872756958, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.015897583216428757, + "learning_rate": 0.000559585254906272, + "loss": 4.377284049987793, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.016513120383024216, + "learning_rate": 0.000559450892376025, + "loss": 4.41200065612793, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.016625918447971344, + "learning_rate": 0.000559316323227926, + "loss": 4.381187438964844, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.016028188169002533, + "learning_rate": 0.000559181547570395, + "loss": 4.353909015655518, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01682914048433304, + "learning_rate": 0.0005590465655120183, + "loss": 4.382326126098633, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.017078889533877373, + "learning_rate": 0.0005589113771615486, + "loss": 4.334949493408203, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.018182938918471336, + "learning_rate": 0.0005587759826279046, + "loss": 4.348130226135254, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01940886303782463, + "learning_rate": 0.0005586403820201713, + "loss": 4.339404106140137, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.020644692704081535, + "learning_rate": 0.0005585045754475995, + "loss": 4.358404159545898, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02433430217206478, + "learning_rate": 0.0005583685630196064, + "loss": 4.3660502433776855, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.027824537828564644, + "learning_rate": 0.0005582323448457744, + "loss": 4.34050178527832, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.02973582223057747, + "learning_rate": 0.0005580959210358523, + "loss": 4.348322868347168, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.030396128073334694, + "learning_rate": 0.0005579592916997542, + "loss": 4.356224060058594, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.029760530218482018, + "learning_rate": 0.0005578224569475599, + "loss": 4.357470989227295, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.02855013869702816, + "learning_rate": 0.0005576854168895147, + "loss": 4.335353374481201, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03376635164022446, + "learning_rate": 0.0005575481716360292, + "loss": 4.331794738769531, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.034431129693984985, + "learning_rate": 0.0005574107212976796, + "loss": 4.336880207061768, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.03524135425686836, + "learning_rate": 0.000557273065985207, + "loss": 4.31537389755249, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.03152058273553848, + "learning_rate": 0.0005571352058095179, + "loss": 4.37662410736084, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02878640592098236, + "learning_rate": 0.0005569971408816838, + "loss": 4.292388916015625, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.029663654044270515, + "learning_rate": 0.0005568588713129409, + "loss": 4.344346523284912, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.027895759791135788, + "learning_rate": 0.0005567203972146906, + "loss": 4.365072727203369, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.026465769857168198, + "learning_rate": 0.000556581718698499, + "loss": 4.334959506988525, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.03191852569580078, + "learning_rate": 0.000556442835876097, + "loss": 4.361069679260254, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.03535444661974907, + "learning_rate": 0.0005563037488593799, + "loss": 4.349765777587891, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03451581299304962, + "learning_rate": 0.0005561644577604074, + "loss": 4.343142509460449, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029810845851898193, + "learning_rate": 0.000556024962691404, + "loss": 4.307485580444336, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.027523217722773552, + "learning_rate": 0.0005558852637647584, + "loss": 4.366406440734863, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03043208085000515, + "learning_rate": 0.0005557453610930235, + "loss": 4.356929779052734, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.03058922104537487, + "learning_rate": 0.0005556052547889161, + "loss": 4.348942279815674, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.032052166759967804, + "learning_rate": 0.0005554649449653177, + "loss": 4.3496809005737305, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.03170694038271904, + "learning_rate": 0.0005553244317352731, + "loss": 4.315959930419922, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 4.4757890701293945, + "eval_runtime": 57.7451, + "eval_samples_per_second": 42.289, + "eval_steps_per_second": 1.333, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.029025450348854065, + "learning_rate": 0.0005551837152119915, + "loss": 4.288647174835205, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.029403962194919586, + "learning_rate": 0.0005550427955088455, + "loss": 4.288555145263672, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.03265019878745079, + "learning_rate": 0.0005549016727393715, + "loss": 4.292625904083252, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.0387202650308609, + "learning_rate": 0.0005547603470172697, + "loss": 4.238926887512207, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.04554206505417824, + "learning_rate": 0.0005546188184564036, + "loss": 4.29966402053833, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.036231476813554764, + "learning_rate": 0.0005544770871708004, + "loss": 4.304398536682129, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.038703009486198425, + "learning_rate": 0.0005543351532746501, + "loss": 4.2945356369018555, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03224619850516319, + "learning_rate": 0.0005541930168823067, + "loss": 4.261351585388184, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.029722856357693672, + "learning_rate": 0.0005540506781082864, + "loss": 4.303143501281738, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.02764045260846615, + "learning_rate": 0.0005539081370672695, + "loss": 4.243679523468018, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.0312200877815485, + "learning_rate": 0.0005537653938740984, + "loss": 4.279904365539551, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.027349065989255905, + "learning_rate": 0.0005536224486437788, + "loss": 4.270228385925293, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.030385233461856842, + "learning_rate": 0.0005534793014914793, + "loss": 4.301590442657471, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.03292274847626686, + "learning_rate": 0.0005533359525325307, + "loss": 4.301828384399414, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.03265977278351784, + "learning_rate": 0.0005531924018824266, + "loss": 4.286200523376465, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.03177836537361145, + "learning_rate": 0.0005530486496568236, + "loss": 4.296319007873535, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.028879934921860695, + "learning_rate": 0.0005529046959715397, + "loss": 4.261706829071045, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.02855812944471836, + "learning_rate": 0.0005527605409425562, + "loss": 4.273903846740723, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03037901036441326, + "learning_rate": 0.0005526161846860161, + "loss": 4.248080253601074, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.029279105365276337, + "learning_rate": 0.0005524716273182245, + "loss": 4.281311988830566, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.02592580020427704, + "learning_rate": 0.0005523268689556489, + "loss": 4.204848289489746, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.021899603307247162, + "learning_rate": 0.0005521819097149183, + "loss": 4.232453346252441, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.019704442471265793, + "learning_rate": 0.0005520367497128238, + "loss": 4.253462314605713, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.019550519064068794, + "learning_rate": 0.0005518913890663182, + "loss": 4.307737350463867, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.020320113748311996, + "learning_rate": 0.0005517458278925161, + "loss": 4.257411956787109, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02118559181690216, + "learning_rate": 0.0005516000663086932, + "loss": 4.262089729309082, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.02050616219639778, + "learning_rate": 0.0005514541044322872, + "loss": 4.2490644454956055, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019793255254626274, + "learning_rate": 0.0005513079423808967, + "loss": 4.267400741577148, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.020760254934430122, + "learning_rate": 0.0005511615802722823, + "loss": 4.2532429695129395, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.019524799659848213, + "learning_rate": 0.0005510150182243647, + "loss": 4.267422199249268, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.020166637375950813, + "learning_rate": 0.0005508682563552269, + "loss": 4.214296817779541, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.01964394748210907, + "learning_rate": 0.0005507212947831119, + "loss": 4.252087116241455, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01690433733165264, + "learning_rate": 0.0005505741336264239, + "loss": 4.232336044311523, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.017682762816548347, + "learning_rate": 0.0005504267730037283, + "loss": 4.213962554931641, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01688133366405964, + "learning_rate": 0.0005502792130337508, + "loss": 4.199461460113525, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.019166061654686928, + "learning_rate": 0.0005501314538353776, + "loss": 4.228936672210693, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.024065518751740456, + "learning_rate": 0.000549983495527656, + "loss": 4.214227676391602, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02518288791179657, + "learning_rate": 0.000549835338229793, + "loss": 4.256412029266357, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.026093967258930206, + "learning_rate": 0.0005496869820611563, + "loss": 4.216564178466797, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.023775635287165642, + "learning_rate": 0.0005495384271412739, + "loss": 4.226430892944336, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.018810907378792763, + "learning_rate": 0.0005493896735898338, + "loss": 4.190013885498047, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.018974872305989265, + "learning_rate": 0.000549240721526684, + "loss": 4.246231555938721, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.02069833129644394, + "learning_rate": 0.0005490915710718325, + "loss": 4.198457717895508, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.02097306028008461, + "learning_rate": 0.0005489422223454471, + "loss": 4.212507247924805, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022165542468428612, + "learning_rate": 0.0005487926754678556, + "loss": 4.2043843269348145, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.019068919122219086, + "learning_rate": 0.0005486429305595448, + "loss": 4.197460651397705, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01701812446117401, + "learning_rate": 0.0005484929877411618, + "loss": 4.220056533813477, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01629726029932499, + "learning_rate": 0.0005483428471335127, + "loss": 4.214868545532227, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018661806359887123, + "learning_rate": 0.000548192508857563, + "loss": 4.227047920227051, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.02005891129374504, + "learning_rate": 0.0005480419730344377, + "loss": 4.197265625, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.02591128647327423, + "learning_rate": 0.0005478912397854208, + "loss": 4.211041450500488, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.032659754157066345, + "learning_rate": 0.0005477403092319552, + "loss": 4.230195045471191, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.04057808965444565, + "learning_rate": 0.0005475891814956429, + "loss": 4.223843097686768, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.04103454574942589, + "learning_rate": 0.0005474378566982449, + "loss": 4.197145462036133, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.03810729831457138, + "learning_rate": 0.0005472863349616809, + "loss": 4.237957000732422, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.028838301077485085, + "learning_rate": 0.0005471346164080292, + "loss": 4.206722259521484, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02436191774904728, + "learning_rate": 0.0005469827011595266, + "loss": 4.1983137130737305, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.022839387878775597, + "learning_rate": 0.0005468305893385686, + "loss": 4.20530366897583, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.02375260926783085, + "learning_rate": 0.0005466782810677088, + "loss": 4.213202476501465, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.02177741937339306, + "learning_rate": 0.0005465257764696593, + "loss": 4.200539588928223, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.02351468615233898, + "learning_rate": 0.0005463730756672903, + "loss": 4.198091506958008, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.023375794291496277, + "learning_rate": 0.0005462201787836301, + "loss": 4.193244934082031, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.02479805052280426, + "learning_rate": 0.0005460670859418651, + "loss": 4.2250871658325195, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.026894841343164444, + "learning_rate": 0.000545913797265339, + "loss": 4.241791248321533, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.030582403764128685, + "learning_rate": 0.0005457603128775541, + "loss": 4.203666687011719, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.032222259789705276, + "learning_rate": 0.0005456066329021699, + "loss": 4.209629535675049, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.03181060031056404, + "learning_rate": 0.0005454527574630037, + "loss": 4.200459957122803, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02687576413154602, + "learning_rate": 0.0005452986866840298, + "loss": 4.211819648742676, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.025988668203353882, + "learning_rate": 0.0005451444206893806, + "loss": 4.198040962219238, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.02937682718038559, + "learning_rate": 0.0005449899596033453, + "loss": 4.222935676574707, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.03043951839208603, + "learning_rate": 0.0005448353035503704, + "loss": 4.2031755447387695, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.030401621013879776, + "learning_rate": 0.0005446804526550596, + "loss": 4.198380470275879, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 4.368444919586182, + "eval_runtime": 58.3009, + "eval_samples_per_second": 41.886, + "eval_steps_per_second": 1.321, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.03461585193872452, + "learning_rate": 0.0005445254070421732, + "loss": 4.147369384765625, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.039084456861019135, + "learning_rate": 0.0005443701668366288, + "loss": 4.124129295349121, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.04021848738193512, + "learning_rate": 0.0005442147321635007, + "loss": 4.127362251281738, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.041000768542289734, + "learning_rate": 0.00054405910314802, + "loss": 4.1432061195373535, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.04008613899350166, + "learning_rate": 0.0005439032799155737, + "loss": 4.169931888580322, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.039394307881593704, + "learning_rate": 0.000543747262591706, + "loss": 4.155813217163086, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.04374868795275688, + "learning_rate": 0.0005435910513021172, + "loss": 4.159891128540039, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0443582646548748, + "learning_rate": 0.0005434346461726641, + "loss": 4.157362937927246, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.04601646214723587, + "learning_rate": 0.0005432780473293593, + "loss": 4.152097225189209, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.04327242821455002, + "learning_rate": 0.0005431212548983718, + "loss": 4.170074462890625, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.0413462370634079, + "learning_rate": 0.0005429642690060262, + "loss": 4.156710147857666, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.036855220794677734, + "learning_rate": 0.0005428070897788034, + "loss": 4.138053894042969, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.03781217709183693, + "learning_rate": 0.0005426497173433398, + "loss": 4.172306060791016, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.03886301815509796, + "learning_rate": 0.0005424921518264275, + "loss": 4.135544776916504, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.03730994835495949, + "learning_rate": 0.0005423343933550143, + "loss": 4.171654224395752, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.03271950036287308, + "learning_rate": 0.0005421764420562032, + "loss": 4.169906139373779, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.03292673081159592, + "learning_rate": 0.0005420182980572527, + "loss": 4.163660049438477, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.029366906732320786, + "learning_rate": 0.0005418599614855768, + "loss": 4.161055088043213, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.023805437609553337, + "learning_rate": 0.0005417014324687442, + "loss": 4.172390937805176, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.023178137838840485, + "learning_rate": 0.000541542711134479, + "loss": 4.146788597106934, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.02403450198471546, + "learning_rate": 0.0005413837976106599, + "loss": 4.124364852905273, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.025009717792272568, + "learning_rate": 0.0005412246920253208, + "loss": 4.12275505065918, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.023431289941072464, + "learning_rate": 0.0005410653945066503, + "loss": 4.134739398956299, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.023803815245628357, + "learning_rate": 0.0005409059051829914, + "loss": 4.130463123321533, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.021219994872808456, + "learning_rate": 0.0005407462241828417, + "loss": 4.124577522277832, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.01899666152894497, + "learning_rate": 0.0005405863516348535, + "loss": 4.167086124420166, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.01889493502676487, + "learning_rate": 0.000540426287667833, + "loss": 4.152744293212891, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.019546914845705032, + "learning_rate": 0.000540266032410741, + "loss": 4.122655868530273, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02085448056459427, + "learning_rate": 0.0005401055859926923, + "loss": 4.141949653625488, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.019959403201937675, + "learning_rate": 0.0005399449485429555, + "loss": 4.131907939910889, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.01885392889380455, + "learning_rate": 0.0005397841201909535, + "loss": 4.118953227996826, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.018216131255030632, + "learning_rate": 0.0005396231010662627, + "loss": 4.096459865570068, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.019385207444429398, + "learning_rate": 0.0005394618912986132, + "loss": 4.111050128936768, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.018676603212952614, + "learning_rate": 0.0005393004910178889, + "loss": 4.156215667724609, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.01876339502632618, + "learning_rate": 0.0005391389003541271, + "loss": 4.1107282638549805, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.01831623539328575, + "learning_rate": 0.0005389771194375185, + "loss": 4.154116630554199, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.018896661698818207, + "learning_rate": 0.0005388151483984071, + "loss": 4.093320846557617, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.019582601264119148, + "learning_rate": 0.0005386529873672898, + "loss": 4.106500625610352, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.020313601940870285, + "learning_rate": 0.0005384906364748172, + "loss": 4.071961879730225, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.020385954529047012, + "learning_rate": 0.000538328095851792, + "loss": 4.129006862640381, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02231123298406601, + "learning_rate": 0.0005381653656291707, + "loss": 4.067756652832031, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.022957179695367813, + "learning_rate": 0.0005380024459380619, + "loss": 4.088230133056641, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021438462659716606, + "learning_rate": 0.000537839336909727, + "loss": 4.128036022186279, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01893841102719307, + "learning_rate": 0.00053767603867558, + "loss": 4.155407905578613, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.018539676442742348, + "learning_rate": 0.0005375125513671874, + "loss": 4.101880073547363, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.01881185919046402, + "learning_rate": 0.0005373488751162681, + "loss": 4.118906021118164, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.021657396107912064, + "learning_rate": 0.0005371850100546929, + "loss": 4.091935157775879, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.023293187841773033, + "learning_rate": 0.000537020956314485, + "loss": 4.105803489685059, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.021306317299604416, + "learning_rate": 0.0005368567140278196, + "loss": 4.0610809326171875, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.022078553214669228, + "learning_rate": 0.0005366922833270235, + "loss": 4.09867525100708, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.02266664244234562, + "learning_rate": 0.0005365276643445757, + "loss": 4.1027326583862305, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.02091757394373417, + "learning_rate": 0.0005363628572131069, + "loss": 4.080266952514648, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.020374659448862076, + "learning_rate": 0.0005361978620653991, + "loss": 4.088479042053223, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.02294750325381756, + "learning_rate": 0.0005360326790343858, + "loss": 4.084364891052246, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.02337353676557541, + "learning_rate": 0.000535867308253152, + "loss": 4.125696182250977, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.022643696516752243, + "learning_rate": 0.0005357017498549341, + "loss": 4.097160339355469, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.021080605685710907, + "learning_rate": 0.0005355360039731196, + "loss": 4.102363109588623, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.02269427664577961, + "learning_rate": 0.0005353700707412466, + "loss": 4.089844703674316, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.02195413038134575, + "learning_rate": 0.0005352039502930048, + "loss": 4.132293701171875, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.01805616170167923, + "learning_rate": 0.0005350376427622343, + "loss": 4.0842814445495605, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.018048783764243126, + "learning_rate": 0.0005348711482829261, + "loss": 4.109591007232666, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.0199508648365736, + "learning_rate": 0.000534704466989222, + "loss": 4.0908074378967285, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.020186785608530045, + "learning_rate": 0.0005345375990154138, + "loss": 4.077959060668945, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.022526193410158157, + "learning_rate": 0.0005343705444959441, + "loss": 4.109979629516602, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.025224441662430763, + "learning_rate": 0.000534203303565406, + "loss": 4.047533988952637, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.027292724698781967, + "learning_rate": 0.000534035876358542, + "loss": 4.09566593170166, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.025669090449810028, + "learning_rate": 0.0005338682630102454, + "loss": 4.0606160163879395, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02791150100529194, + "learning_rate": 0.0005337004636555595, + "loss": 4.097081184387207, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.03294133022427559, + "learning_rate": 0.0005335324784296768, + "loss": 4.121442794799805, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.027322066947817802, + "learning_rate": 0.0005333643074679403, + "loss": 4.1188740730285645, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.024866372346878052, + "learning_rate": 0.000533195950905842, + "loss": 4.082042694091797, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.02901027724146843, + "learning_rate": 0.0005330274088790239, + "loss": 4.110199451446533, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 4.283970355987549, + "eval_runtime": 58.4003, + "eval_samples_per_second": 41.815, + "eval_steps_per_second": 1.318, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.028958842158317566, + "learning_rate": 0.0005328586815232773, + "loss": 4.040733337402344, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.03491180017590523, + "learning_rate": 0.0005326897689745428, + "loss": 4.0473432540893555, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.04025818780064583, + "learning_rate": 0.0005325206713689101, + "loss": 4.021976470947266, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03998899459838867, + "learning_rate": 0.0005323513888426181, + "loss": 4.04901123046875, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.042165111750364304, + "learning_rate": 0.0005321819215320546, + "loss": 4.050230979919434, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.03511597216129303, + "learning_rate": 0.0005320122695737565, + "loss": 4.0989990234375, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.04064979776740074, + "learning_rate": 0.0005318424331044094, + "loss": 4.051448822021484, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.047536592930555344, + "learning_rate": 0.0005316724122608473, + "loss": 4.064278602600098, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.05592669919133186, + "learning_rate": 0.0005315022071800527, + "loss": 4.045516014099121, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.04262633994221687, + "learning_rate": 0.0005313318179991574, + "loss": 4.060666084289551, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03997781500220299, + "learning_rate": 0.0005311612448554402, + "loss": 4.065315246582031, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.038152433931827545, + "learning_rate": 0.0005309904878863293, + "loss": 4.069721221923828, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.0367623008787632, + "learning_rate": 0.0005308195472294001, + "loss": 4.080986976623535, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.0358598455786705, + "learning_rate": 0.0005306484230223769, + "loss": 4.065290927886963, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.039188165217638016, + "learning_rate": 0.000530477115403131, + "loss": 4.024333477020264, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.036839909851551056, + "learning_rate": 0.0005303056245096817, + "loss": 4.067532539367676, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.03456985577940941, + "learning_rate": 0.0005301339504801967, + "loss": 4.048443794250488, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.039487943053245544, + "learning_rate": 0.0005299620934529902, + "loss": 4.0585737228393555, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.03415827825665474, + "learning_rate": 0.0005297900535665244, + "loss": 4.06104850769043, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.03209620714187622, + "learning_rate": 0.0005296178309594087, + "loss": 4.009165287017822, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.028947360813617706, + "learning_rate": 0.0005294454257704, + "loss": 4.067384243011475, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026836825534701347, + "learning_rate": 0.0005292728381384018, + "loss": 4.046419620513916, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024095216765999794, + "learning_rate": 0.0005291000682024647, + "loss": 4.019008636474609, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.020965851843357086, + "learning_rate": 0.0005289271161017867, + "loss": 4.020613670349121, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.02250855043530464, + "learning_rate": 0.0005287539819757119, + "loss": 4.031634330749512, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.021385207772254944, + "learning_rate": 0.0005285806659637315, + "loss": 4.086246013641357, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.021104589104652405, + "learning_rate": 0.000528407168205483, + "loss": 4.033970355987549, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.019584037363529205, + "learning_rate": 0.0005282334888407503, + "loss": 4.025118827819824, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.02107258513569832, + "learning_rate": 0.000528059628009464, + "loss": 4.020823955535889, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.020438389852643013, + "learning_rate": 0.0005278855858517003, + "loss": 4.016323566436768, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.01881488971412182, + "learning_rate": 0.0005277113625076822, + "loss": 3.9991841316223145, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.01763618178665638, + "learning_rate": 0.000527536958117778, + "loss": 4.0062150955200195, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.017308862879872322, + "learning_rate": 0.0005273623728225022, + "loss": 3.946723461151123, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01706845872104168, + "learning_rate": 0.0005271876067625153, + "loss": 4.056562900543213, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.018664294853806496, + "learning_rate": 0.0005270126600786228, + "loss": 4.024652481079102, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.018523404374718666, + "learning_rate": 0.0005268375329117762, + "loss": 3.962064743041992, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.017408935353159904, + "learning_rate": 0.0005266622254030724, + "loss": 4.008697509765625, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.01588570512831211, + "learning_rate": 0.0005264867376937534, + "loss": 4.011664390563965, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.015084310434758663, + "learning_rate": 0.0005263110699252065, + "loss": 3.996224880218506, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01637054979801178, + "learning_rate": 0.000526135222238964, + "loss": 4.015972137451172, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.01673802360892296, + "learning_rate": 0.0005259591947767032, + "loss": 3.975414276123047, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.016544019803404808, + "learning_rate": 0.0005257829876802464, + "loss": 4.012955188751221, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.017455460503697395, + "learning_rate": 0.0005256066010915603, + "loss": 4.010180473327637, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.017625460401177406, + "learning_rate": 0.0005254300351527566, + "loss": 4.001679420471191, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.01732795126736164, + "learning_rate": 0.000525253290006091, + "loss": 4.006107330322266, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.016654646024107933, + "learning_rate": 0.0005250763657939641, + "loss": 4.045180797576904, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.015085658989846706, + "learning_rate": 0.0005248992626589203, + "loss": 3.994246482849121, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.015112568624317646, + "learning_rate": 0.0005247219807436489, + "loss": 4.004963397979736, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.014882124960422516, + "learning_rate": 0.000524544520190982, + "loss": 4.029565811157227, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.016012471169233322, + "learning_rate": 0.000524366881143897, + "loss": 4.018208026885986, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.015761924907565117, + "learning_rate": 0.0005241890637455141, + "loss": 3.9911205768585205, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.017770355567336082, + "learning_rate": 0.0005240110681390978, + "loss": 4.028733253479004, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01959751546382904, + "learning_rate": 0.0005238328944680558, + "loss": 4.003322124481201, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.018713824450969696, + "learning_rate": 0.0005236545428759394, + "loss": 3.991718292236328, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.01992208883166313, + "learning_rate": 0.0005234760135064434, + "loss": 4.002507209777832, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.021671291440725327, + "learning_rate": 0.0005232973065034055, + "loss": 4.021344184875488, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024282798171043396, + "learning_rate": 0.0005231184220108069, + "loss": 3.9899158477783203, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.023294344544410706, + "learning_rate": 0.0005229393601727713, + "loss": 4.046627044677734, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.02029707282781601, + "learning_rate": 0.000522760121133566, + "loss": 4.022454261779785, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.02151239849627018, + "learning_rate": 0.0005225807050376004, + "loss": 4.005476951599121, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.02320881187915802, + "learning_rate": 0.0005224011120294269, + "loss": 4.012726783752441, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.026484191417694092, + "learning_rate": 0.0005222213422537401, + "loss": 3.9904391765594482, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.026279661804437637, + "learning_rate": 0.0005220413958553777, + "loss": 3.9738848209381104, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.024299729615449905, + "learning_rate": 0.0005218612729793188, + "loss": 4.003316402435303, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.02337595447897911, + "learning_rate": 0.0005216809737706856, + "loss": 3.954984188079834, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.02416428178548813, + "learning_rate": 0.0005215004983747415, + "loss": 4.0317063331604, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.022488107904791832, + "learning_rate": 0.0005213198469368925, + "loss": 3.9736647605895996, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.019700225442647934, + "learning_rate": 0.0005211390196026861, + "loss": 3.979867458343506, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.020382333546876907, + "learning_rate": 0.0005209580165178117, + "loss": 3.995251178741455, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.020149298012256622, + "learning_rate": 0.0005207768378281003, + "loss": 3.9987268447875977, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.019790509715676308, + "learning_rate": 0.0005205954836795241, + "loss": 3.9838991165161133, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.022096451371908188, + "learning_rate": 0.000520413954218197, + "loss": 3.995800256729126, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.214347839355469, + "eval_runtime": 58.8872, + "eval_samples_per_second": 41.469, + "eval_steps_per_second": 1.308, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02203402668237686, + "learning_rate": 0.0005202322495903739, + "loss": 3.915332078933716, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.02336675301194191, + "learning_rate": 0.0005200503699424509, + "loss": 3.914801597595215, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.027339637279510498, + "learning_rate": 0.0005198683154209654, + "loss": 3.9277780055999756, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.026872653514146805, + "learning_rate": 0.0005196860861725952, + "loss": 3.9160094261169434, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.026558255776762962, + "learning_rate": 0.0005195036823441592, + "loss": 3.9084300994873047, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.02644367143511772, + "learning_rate": 0.0005193211040826169, + "loss": 3.9328718185424805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.024677276611328125, + "learning_rate": 0.0005191383515350681, + "loss": 3.88993501663208, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.023194244131445885, + "learning_rate": 0.0005189554248487536, + "loss": 3.923567056655884, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.024653147906064987, + "learning_rate": 0.0005187723241710539, + "loss": 3.8740999698638916, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.026041202247142792, + "learning_rate": 0.0005185890496494899, + "loss": 3.9542148113250732, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.029586758464574814, + "learning_rate": 0.0005184056014317225, + "loss": 3.9241347312927246, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.0325787179172039, + "learning_rate": 0.0005182219796655528, + "loss": 3.9656593799591064, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.03183380514383316, + "learning_rate": 0.0005180381844989214, + "loss": 3.950648307800293, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.03434471786022186, + "learning_rate": 0.0005178542160799088, + "loss": 3.9590439796447754, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.030099432915449142, + "learning_rate": 0.0005176700745567353, + "loss": 3.9518113136291504, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.025950564071536064, + "learning_rate": 0.0005174857600777599, + "loss": 3.8908939361572266, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.029959935694932938, + "learning_rate": 0.0005173012727914816, + "loss": 3.945037364959717, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.029551763087511063, + "learning_rate": 0.0005171166128465389, + "loss": 3.923959732055664, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.03538523241877556, + "learning_rate": 0.0005169317803917083, + "loss": 3.9395527839660645, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.037924155592918396, + "learning_rate": 0.0005167467755759065, + "loss": 3.945115089416504, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.036773014813661575, + "learning_rate": 0.0005165615985481881, + "loss": 3.9484646320343018, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.04125838354229927, + "learning_rate": 0.0005163762494577474, + "loss": 4.000699043273926, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.044125135987997055, + "learning_rate": 0.0005161907284539165, + "loss": 3.9828028678894043, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.040853384882211685, + "learning_rate": 0.0005160050356861663, + "loss": 3.957505702972412, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.03501008450984955, + "learning_rate": 0.0005158191713041063, + "loss": 3.9360849857330322, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.04082556813955307, + "learning_rate": 0.0005156331354574838, + "loss": 3.95941424369812, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.045432962477207184, + "learning_rate": 0.0005154469282961847, + "loss": 3.9666922092437744, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.04048244655132294, + "learning_rate": 0.0005152605499702328, + "loss": 3.984555244445801, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.031016314402222633, + "learning_rate": 0.0005150740006297896, + "loss": 3.945204257965088, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.03210771828889847, + "learning_rate": 0.0005148872804251548, + "loss": 3.976186752319336, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.03130023553967476, + "learning_rate": 0.000514700389506765, + "loss": 3.915727138519287, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.031193193048238754, + "learning_rate": 0.0005145133280251954, + "loss": 3.961866855621338, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.027518954128026962, + "learning_rate": 0.0005143260961311578, + "loss": 3.9742956161499023, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.02751155197620392, + "learning_rate": 0.0005141386939755014, + "loss": 3.949662208557129, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.026033353060483932, + "learning_rate": 0.0005139511217092129, + "loss": 3.9327926635742188, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.02579621411859989, + "learning_rate": 0.000513763379483416, + "loss": 3.9334418773651123, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.023009559139609337, + "learning_rate": 0.0005135754674493709, + "loss": 3.969707489013672, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.02196447364985943, + "learning_rate": 0.0005133873857584753, + "loss": 3.9084434509277344, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.02206484042108059, + "learning_rate": 0.0005131991345622629, + "loss": 3.922117233276367, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.023360205814242363, + "learning_rate": 0.0005130107140124046, + "loss": 3.965158224105835, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02162669226527214, + "learning_rate": 0.0005128221242607072, + "loss": 3.9519729614257812, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.019719846546649933, + "learning_rate": 0.0005126333654591141, + "loss": 3.9195961952209473, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.018988260999321938, + "learning_rate": 0.0005124444377597049, + "loss": 3.937926769256592, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.019278401508927345, + "learning_rate": 0.0005122553413146951, + "loss": 3.9126808643341064, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019101005047559738, + "learning_rate": 0.0005120660762764366, + "loss": 3.8993749618530273, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.019156700000166893, + "learning_rate": 0.0005118766427974166, + "loss": 3.9531068801879883, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019944407045841217, + "learning_rate": 0.0005116870410302582, + "loss": 3.9645180702209473, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.01826174184679985, + "learning_rate": 0.0005114972711277202, + "loss": 3.948477268218994, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.0191368218511343, + "learning_rate": 0.0005113073332426968, + "loss": 3.944648027420044, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.017768343910574913, + "learning_rate": 0.0005111172275282174, + "loss": 3.9584474563598633, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.01955590397119522, + "learning_rate": 0.0005109269541374469, + "loss": 3.8936800956726074, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.02100536972284317, + "learning_rate": 0.000510736513223685, + "loss": 3.93919038772583, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.020873935893177986, + "learning_rate": 0.0005105459049403665, + "loss": 3.929244041442871, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.021876046434044838, + "learning_rate": 0.0005103551294410614, + "loss": 3.9264731407165527, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.021726684644818306, + "learning_rate": 0.0005101641868794736, + "loss": 3.923135280609131, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.019308188930153847, + "learning_rate": 0.0005099730774094423, + "loss": 3.9062135219573975, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01917017623782158, + "learning_rate": 0.0005097818011849408, + "loss": 3.898810386657715, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017933931201696396, + "learning_rate": 0.0005095903583600772, + "loss": 3.8899734020233154, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.015131795778870583, + "learning_rate": 0.0005093987490890933, + "loss": 3.9452900886535645, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.017190825194120407, + "learning_rate": 0.0005092069735263651, + "loss": 3.9124746322631836, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.01618487387895584, + "learning_rate": 0.000509015031826403, + "loss": 3.91654109954834, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.014920198358595371, + "learning_rate": 0.0005088229241438509, + "loss": 3.9144535064697266, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01615205965936184, + "learning_rate": 0.0005086306506334864, + "loss": 3.917003631591797, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.015370192006230354, + "learning_rate": 0.0005084382114502208, + "loss": 3.8910603523254395, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.015290996059775352, + "learning_rate": 0.0005082456067490988, + "loss": 3.9122695922851562, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.015913106501102448, + "learning_rate": 0.0005080528366852987, + "loss": 3.9304556846618652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.01639600843191147, + "learning_rate": 0.0005078599014141318, + "loss": 3.8974502086639404, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.017010601237416267, + "learning_rate": 0.0005076668010910425, + "loss": 3.9723148345947266, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.017778996378183365, + "learning_rate": 0.0005074735358716083, + "loss": 3.8905391693115234, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.020029611885547638, + "learning_rate": 0.0005072801059115394, + "loss": 3.934976100921631, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.01931857503950596, + "learning_rate": 0.000507086511366679, + "loss": 3.9152774810791016, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019750485196709633, + "learning_rate": 0.0005068927523930027, + "loss": 3.877650499343872, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.161914825439453, + "eval_runtime": 58.5217, + "eval_samples_per_second": 41.728, + "eval_steps_per_second": 1.316, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02073049545288086, + "learning_rate": 0.0005066988291466184, + "loss": 3.8331246376037598, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02272389829158783, + "learning_rate": 0.0005065047417837669, + "loss": 3.8847155570983887, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02139078453183174, + "learning_rate": 0.0005063104904608206, + "loss": 3.8276169300079346, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02043229155242443, + "learning_rate": 0.0005061160753342843, + "loss": 3.8804001808166504, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.02345564216375351, + "learning_rate": 0.0005059214965607948, + "loss": 3.851609468460083, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.023448185995221138, + "learning_rate": 0.0005057267542971207, + "loss": 3.8446507453918457, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.024826066568493843, + "learning_rate": 0.0005055318487001624, + "loss": 3.86712384223938, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.02491980977356434, + "learning_rate": 0.0005053367799269518, + "loss": 3.8374528884887695, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.023002834990620613, + "learning_rate": 0.0005051415481346522, + "loss": 3.783214569091797, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.02123952843248844, + "learning_rate": 0.0005049461534805585, + "loss": 3.8463428020477295, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.01981225423514843, + "learning_rate": 0.0005047505961220964, + "loss": 3.8307082653045654, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01915367692708969, + "learning_rate": 0.0005045548762168231, + "loss": 3.8190226554870605, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.01890859752893448, + "learning_rate": 0.0005043589939224266, + "loss": 3.8428220748901367, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.019150376319885254, + "learning_rate": 0.0005041629493967257, + "loss": 3.830247402191162, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.020099453628063202, + "learning_rate": 0.0005039667427976697, + "loss": 3.838806629180908, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.021136781200766563, + "learning_rate": 0.0005037703742833391, + "loss": 3.8344273567199707, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.022857150062918663, + "learning_rate": 0.0005035738440119442, + "loss": 3.8317880630493164, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.024891739711165428, + "learning_rate": 0.0005033771521418261, + "loss": 3.837855577468872, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.02595464698970318, + "learning_rate": 0.0005031802988314557, + "loss": 3.846050262451172, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.02722409926354885, + "learning_rate": 0.0005029832842394342, + "loss": 3.8164305686950684, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.024486860260367393, + "learning_rate": 0.0005027861085244927, + "loss": 3.8099875450134277, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.021551169455051422, + "learning_rate": 0.0005025887718454922, + "loss": 3.825925827026367, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.02121506817638874, + "learning_rate": 0.0005023912743614232, + "loss": 3.870727777481079, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.01948230341076851, + "learning_rate": 0.0005021936162314058, + "loss": 3.8035314083099365, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.01896044984459877, + "learning_rate": 0.0005019957976146898, + "loss": 3.84665584564209, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.020620400086045265, + "learning_rate": 0.0005017978186706538, + "loss": 3.8338310718536377, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.020855262875556946, + "learning_rate": 0.0005015996795588062, + "loss": 3.8781003952026367, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.02153870463371277, + "learning_rate": 0.0005014013804387838, + "loss": 3.8783490657806396, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.023751024156808853, + "learning_rate": 0.0005012029214703529, + "loss": 3.846198558807373, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.023987066000699997, + "learning_rate": 0.000501004302813408, + "loss": 3.834296226501465, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.023626182228326797, + "learning_rate": 0.0005008055246279727, + "loss": 3.8591227531433105, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.025707192718982697, + "learning_rate": 0.000500606587074199, + "loss": 3.8766489028930664, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.027723029255867004, + "learning_rate": 0.0005004074903123674, + "loss": 3.88264799118042, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.0278895553201437, + "learning_rate": 0.0005002082345028864, + "loss": 3.8932995796203613, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.025622662156820297, + "learning_rate": 0.0005000088198062929, + "loss": 3.8734612464904785, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.02639615908265114, + "learning_rate": 0.0004998092463832516, + "loss": 3.786813735961914, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.026765599846839905, + "learning_rate": 0.0004996095143945554, + "loss": 3.853966474533081, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02823319099843502, + "learning_rate": 0.0004994096240011246, + "loss": 3.8321340084075928, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.023394955322146416, + "learning_rate": 0.0004992095753640072, + "loss": 3.8527331352233887, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.023874662816524506, + "learning_rate": 0.0004990093686443791, + "loss": 3.8822879791259766, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.02301311306655407, + "learning_rate": 0.000498809004003543, + "loss": 3.846186876296997, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.022630592808127403, + "learning_rate": 0.0004986084816029291, + "loss": 3.867992639541626, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.02424517087638378, + "learning_rate": 0.0004984078016040946, + "loss": 3.867293357849121, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.024116624146699905, + "learning_rate": 0.000498206964168724, + "loss": 3.846282958984375, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.02187768742442131, + "learning_rate": 0.000498005969458628, + "loss": 3.8429484367370605, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.01879809983074665, + "learning_rate": 0.0004978048176357447, + "loss": 3.863215446472168, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.020426234230399132, + "learning_rate": 0.0004976035088621385, + "loss": 3.825230121612549, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01992812193930149, + "learning_rate": 0.0004974020433000003, + "loss": 3.851665496826172, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.020693328231573105, + "learning_rate": 0.000497200421111647, + "loss": 3.906999111175537, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.021137919276952744, + "learning_rate": 0.0004969986424595221, + "loss": 3.8369035720825195, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.023285407572984695, + "learning_rate": 0.0004967967075061949, + "loss": 3.8658390045166016, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.0219135619699955, + "learning_rate": 0.0004965946164143609, + "loss": 3.873072862625122, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.020101983100175858, + "learning_rate": 0.0004963923693468413, + "loss": 3.847805976867676, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.01875176839530468, + "learning_rate": 0.0004961899664665827, + "loss": 3.7976622581481934, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01913360133767128, + "learning_rate": 0.0004959874079366576, + "loss": 3.853851318359375, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.016863619908690453, + "learning_rate": 0.0004957846939202637, + "loss": 3.877378225326538, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.017802366986870766, + "learning_rate": 0.000495581824580724, + "loss": 3.861356019973755, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.01981814205646515, + "learning_rate": 0.0004953788000814866, + "loss": 3.88613224029541, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.017508897930383682, + "learning_rate": 0.000495175620586125, + "loss": 3.8202030658721924, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.019240548834204674, + "learning_rate": 0.0004949722862583368, + "loss": 3.853827714920044, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.022305598482489586, + "learning_rate": 0.000494768797261945, + "loss": 3.885148763656616, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.02531025931239128, + "learning_rate": 0.000494565153760897, + "loss": 3.8412041664123535, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.024474849924445152, + "learning_rate": 0.0004943613559192648, + "loss": 3.865644931793213, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.020430924370884895, + "learning_rate": 0.0004941574039012444, + "loss": 3.8649442195892334, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.01906431093811989, + "learning_rate": 0.0004939532978711566, + "loss": 3.8462204933166504, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.021118057891726494, + "learning_rate": 0.0004937490379934456, + "loss": 3.8652522563934326, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.020337693393230438, + "learning_rate": 0.00049354462443268, + "loss": 3.8506221771240234, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.019090797752141953, + "learning_rate": 0.0004933400573535521, + "loss": 3.881237506866455, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.020973797887563705, + "learning_rate": 0.000493135336920878, + "loss": 3.821274757385254, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.021984362974762917, + "learning_rate": 0.0004929304632995973, + "loss": 3.8793835639953613, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.0215684212744236, + "learning_rate": 0.0004927254366547727, + "loss": 3.8259663581848145, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.021673865616321564, + "learning_rate": 0.0004925202571515908, + "loss": 3.8505711555480957, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.123871326446533, + "eval_runtime": 58.8357, + "eval_samples_per_second": 41.505, + "eval_steps_per_second": 1.309, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.0203359667211771, + "learning_rate": 0.0004923149249553608, + "loss": 3.785353660583496, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.023197868838906288, + "learning_rate": 0.0004921094402315152, + "loss": 3.7479515075683594, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.02314298041164875, + "learning_rate": 0.0004919038031456093, + "loss": 3.7697315216064453, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.0256908368319273, + "learning_rate": 0.0004916980138633213, + "loss": 3.79056978225708, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.026193121448159218, + "learning_rate": 0.0004914920725504518, + "loss": 3.7722532749176025, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.027120303362607956, + "learning_rate": 0.000491285979372924, + "loss": 3.7991981506347656, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.0266975536942482, + "learning_rate": 0.0004910797344967834, + "loss": 3.778301477432251, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.02806396596133709, + "learning_rate": 0.000490873338088198, + "loss": 3.7578704357147217, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.02893839031457901, + "learning_rate": 0.0004906667903134573, + "loss": 3.7401275634765625, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.031170370057225227, + "learning_rate": 0.0004904600913389735, + "loss": 3.780686616897583, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.03364266827702522, + "learning_rate": 0.0004902532413312799, + "loss": 3.794442653656006, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.030121006071567535, + "learning_rate": 0.0004900462404570321, + "loss": 3.779569149017334, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.03296204283833504, + "learning_rate": 0.0004898390888830067, + "loss": 3.7960336208343506, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.03381749242544174, + "learning_rate": 0.0004896317867761022, + "loss": 3.7959280014038086, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.035383883863687515, + "learning_rate": 0.000489424334303338, + "loss": 3.8008625507354736, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.03759761154651642, + "learning_rate": 0.0004892167316318548, + "loss": 3.7997028827667236, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.030651796609163284, + "learning_rate": 0.0004890089789289147, + "loss": 3.81191349029541, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.0262477844953537, + "learning_rate": 0.0004888010763618997, + "loss": 3.8046622276306152, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.030398324131965637, + "learning_rate": 0.0004885930240983136, + "loss": 3.7903637886047363, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.030515259131789207, + "learning_rate": 0.0004883848223057801, + "loss": 3.8111231327056885, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.031686048954725266, + "learning_rate": 0.00048817647115204386, + "loss": 3.781445026397705, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.025968654081225395, + "learning_rate": 0.00048796797080496936, + "loss": 3.788196563720703, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.025340190157294273, + "learning_rate": 0.00048775932143254176, + "loss": 3.763331413269043, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.024738596752285957, + "learning_rate": 0.0004875505232028661, + "loss": 3.8005969524383545, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.021990075707435608, + "learning_rate": 0.0004873415762841672, + "loss": 3.795844078063965, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.02191784791648388, + "learning_rate": 0.00048713248084478986, + "loss": 3.8062384128570557, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.023678259924054146, + "learning_rate": 0.0004869232370531986, + "loss": 3.7760679721832275, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.024137988686561584, + "learning_rate": 0.0004867138450779773, + "loss": 3.7917492389678955, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.02673223614692688, + "learning_rate": 0.0004865043050878293, + "loss": 3.8006649017333984, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.02395169623196125, + "learning_rate": 0.0004862946172515771, + "loss": 3.7902212142944336, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.02410844899713993, + "learning_rate": 0.0004860847817381625, + "loss": 3.8073291778564453, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.028109831735491753, + "learning_rate": 0.0004858747987166463, + "loss": 3.80116605758667, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.025538964197039604, + "learning_rate": 0.00048566466835620774, + "loss": 3.791151523590088, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.02393815852701664, + "learning_rate": 0.0004854543908261453, + "loss": 3.7766106128692627, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.024659741669893265, + "learning_rate": 0.0004852439662958756, + "loss": 3.789135217666626, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.02668776921927929, + "learning_rate": 0.000485033394934934, + "loss": 3.797905445098877, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.0258641317486763, + "learning_rate": 0.000484822676912974, + "loss": 3.783323287963867, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.024846022948622704, + "learning_rate": 0.0004846118123997674, + "loss": 3.7976717948913574, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.024981874972581863, + "learning_rate": 0.00048440080156520387, + "loss": 3.818732738494873, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.02556881681084633, + "learning_rate": 0.0004841896445792909, + "loss": 3.7821896076202393, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.02558571845293045, + "learning_rate": 0.000483978341612154, + "loss": 3.8143680095672607, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.02361566200852394, + "learning_rate": 0.0004837668928340362, + "loss": 3.7710189819335938, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022570716217160225, + "learning_rate": 0.0004835552984152978, + "loss": 3.775564670562744, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.022478004917502403, + "learning_rate": 0.00048334355852641664, + "loss": 3.7920689582824707, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021241728216409683, + "learning_rate": 0.0004831316733379879, + "loss": 3.7857837677001953, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.018921511247754097, + "learning_rate": 0.0004829196430207235, + "loss": 3.7659268379211426, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01926037296652794, + "learning_rate": 0.00048270746774545237, + "loss": 3.7527129650115967, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.018743667751550674, + "learning_rate": 0.0004824951476831204, + "loss": 3.767991065979004, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.017273949459195137, + "learning_rate": 0.00048228268300479015, + "loss": 3.7834174633026123, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.017054244875907898, + "learning_rate": 0.0004820700738816403, + "loss": 3.7775752544403076, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017190931364893913, + "learning_rate": 0.0004818573204849664, + "loss": 3.7709648609161377, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01665034517645836, + "learning_rate": 0.00048164442298617987, + "loss": 3.7575321197509766, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.016932843253016472, + "learning_rate": 0.00048143138155680844, + "loss": 3.7922799587249756, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01702750287950039, + "learning_rate": 0.0004812181963684958, + "loss": 3.7729780673980713, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01766599342226982, + "learning_rate": 0.00048100486759300134, + "loss": 3.7876572608947754, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01785697042942047, + "learning_rate": 0.00048079139540220024, + "loss": 3.8131699562072754, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01830688677728176, + "learning_rate": 0.0004805777799680831, + "loss": 3.790778636932373, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017348438501358032, + "learning_rate": 0.0004803640214627561, + "loss": 3.7984330654144287, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017999377101659775, + "learning_rate": 0.0004801501200584406, + "loss": 3.7756147384643555, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.01801210641860962, + "learning_rate": 0.000479936075927473, + "loss": 3.789759874343872, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.018445398658514023, + "learning_rate": 0.000479721889242305, + "loss": 3.7683000564575195, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.017544370144605637, + "learning_rate": 0.0004795075601755027, + "loss": 3.7776038646698, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.019834233447909355, + "learning_rate": 0.00047929308889974737, + "loss": 3.7559571266174316, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01791626401245594, + "learning_rate": 0.0004790784755878346, + "loss": 3.7746729850769043, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.017638562247157097, + "learning_rate": 0.00047886372041267434, + "loss": 3.7933082580566406, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.018974050879478455, + "learning_rate": 0.0004786488235472909, + "loss": 3.795713424682617, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.018694303929805756, + "learning_rate": 0.0004784337851648231, + "loss": 3.7675955295562744, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.01788092777132988, + "learning_rate": 0.0004782186054385233, + "loss": 3.786083698272705, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.01902497000992298, + "learning_rate": 0.0004780032845417579, + "loss": 3.7976233959198, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.020558880642056465, + "learning_rate": 0.000477787822648007, + "loss": 3.7500152587890625, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.020093899220228195, + "learning_rate": 0.00047757221993086443, + "loss": 3.703448534011841, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.019833799451589584, + "learning_rate": 0.00047735647656403754, + "loss": 3.756059408187866, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.093987941741943, + "eval_runtime": 57.6894, + "eval_samples_per_second": 42.33, + "eval_steps_per_second": 1.335, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.022167526185512543, + "learning_rate": 0.0004771405927213467, + "loss": 3.696444511413574, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.027874166145920753, + "learning_rate": 0.0004769245685767255, + "loss": 3.713630199432373, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.030872784554958344, + "learning_rate": 0.00047670840430422113, + "loss": 3.678295612335205, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.03246859833598137, + "learning_rate": 0.0004764921000779928, + "loss": 3.7186388969421387, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.03667069226503372, + "learning_rate": 0.0004762756560723131, + "loss": 3.7457995414733887, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.036286015063524246, + "learning_rate": 0.0004760590724615671, + "loss": 3.6881775856018066, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.03413614630699158, + "learning_rate": 0.0004758423494202522, + "loss": 3.6999263763427734, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.03384552150964737, + "learning_rate": 0.0004756254871229782, + "loss": 3.6959803104400635, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.03227181360125542, + "learning_rate": 0.00047540848574446724, + "loss": 3.7182722091674805, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.036366235464811325, + "learning_rate": 0.0004751913454595535, + "loss": 3.7116851806640625, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.03573326766490936, + "learning_rate": 0.0004749740664431828, + "loss": 3.741865873336792, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.03482096269726753, + "learning_rate": 0.000474756648870413, + "loss": 3.7195520401000977, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.03232625126838684, + "learning_rate": 0.0004745390929164134, + "loss": 3.7443251609802246, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.03421664610505104, + "learning_rate": 0.0004743213987564651, + "loss": 3.6867361068725586, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.03585013002157211, + "learning_rate": 0.0004741035665659603, + "loss": 3.7492403984069824, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.029985329136252403, + "learning_rate": 0.00047388559652040236, + "loss": 3.705209493637085, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.02729589305818081, + "learning_rate": 0.0004736674887954059, + "loss": 3.711106538772583, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.029642216861248016, + "learning_rate": 0.0004734492435666963, + "loss": 3.7274727821350098, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.02816903032362461, + "learning_rate": 0.0004732308610101099, + "loss": 3.7181742191314697, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.027143625542521477, + "learning_rate": 0.00047301234130159346, + "loss": 3.719235897064209, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.025802744552493095, + "learning_rate": 0.00047279368461720445, + "loss": 3.771820545196533, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.026424532756209373, + "learning_rate": 0.0004725748911331106, + "loss": 3.7516651153564453, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.029677648097276688, + "learning_rate": 0.0004723559610255898, + "loss": 3.7058472633361816, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.031548649072647095, + "learning_rate": 0.0004721368944710302, + "loss": 3.756992816925049, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.027667395770549774, + "learning_rate": 0.00047191769164592974, + "loss": 3.742711067199707, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.027639545500278473, + "learning_rate": 0.000471698352726896, + "loss": 3.728883743286133, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.02950332500040531, + "learning_rate": 0.0004714788778906467, + "loss": 3.7160212993621826, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.026302607730031013, + "learning_rate": 0.0004712592673140084, + "loss": 3.735713481903076, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.022483112290501595, + "learning_rate": 0.00047103952117391764, + "loss": 3.702686309814453, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.021729422733187675, + "learning_rate": 0.00047081963964741986, + "loss": 3.7195816040039062, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.021159254014492035, + "learning_rate": 0.0004705996229116696, + "loss": 3.684321880340576, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.02233424223959446, + "learning_rate": 0.0004703794711439304, + "loss": 3.7233057022094727, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.02142048068344593, + "learning_rate": 0.0004701591845215744, + "loss": 3.7083680629730225, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.022331461310386658, + "learning_rate": 0.00046993876322208274, + "loss": 3.7408461570739746, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.020629866048693657, + "learning_rate": 0.0004697182074230448, + "loss": 3.759129047393799, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.02174685336649418, + "learning_rate": 0.0004694975173021586, + "loss": 3.72351336479187, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.021538110449910164, + "learning_rate": 0.0004692766930372299, + "loss": 3.716416597366333, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.020800625905394554, + "learning_rate": 0.00046905573480617285, + "loss": 3.726954936981201, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.02039588987827301, + "learning_rate": 0.00046883464278700966, + "loss": 3.7336390018463135, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.02107219770550728, + "learning_rate": 0.0004686134171578702, + "loss": 3.722311496734619, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.02068956382572651, + "learning_rate": 0.00046839205809699183, + "loss": 3.767139434814453, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.020190054550766945, + "learning_rate": 0.0004681705657827195, + "loss": 3.6960511207580566, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.018814411014318466, + "learning_rate": 0.0004679489403935058, + "loss": 3.70375919342041, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.017568519338965416, + "learning_rate": 0.00046772718210791005, + "loss": 3.756627321243286, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.018188195303082466, + "learning_rate": 0.00046750529110459903, + "loss": 3.6964683532714844, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.018925746902823448, + "learning_rate": 0.0004672832675623463, + "loss": 3.731172561645508, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.01690756157040596, + "learning_rate": 0.0004670611116600322, + "loss": 3.6955318450927734, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016300756484270096, + "learning_rate": 0.00046683882357664375, + "loss": 3.762632369995117, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.016405068337917328, + "learning_rate": 0.0004666164034912744, + "loss": 3.75130558013916, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.017687877640128136, + "learning_rate": 0.00046639385158312406, + "loss": 3.687530755996704, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.017524344846606255, + "learning_rate": 0.00046617116803149874, + "loss": 3.738790273666382, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.01714804768562317, + "learning_rate": 0.00046594835301581053, + "loss": 3.69827938079834, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018042324110865593, + "learning_rate": 0.00046572540671557745, + "loss": 3.71600341796875, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018648479133844376, + "learning_rate": 0.00046550232931042346, + "loss": 3.742591619491577, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.019198937341570854, + "learning_rate": 0.00046527912098007787, + "loss": 3.739820957183838, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.01796373911201954, + "learning_rate": 0.00046505578190437564, + "loss": 3.719025135040283, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01836620457470417, + "learning_rate": 0.0004648323122632569, + "loss": 3.741434097290039, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.02192627638578415, + "learning_rate": 0.0004646087122367673, + "loss": 3.7553882598876953, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.02147185243666172, + "learning_rate": 0.00046438498200505727, + "loss": 3.7705790996551514, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01855960488319397, + "learning_rate": 0.0004641611217483821, + "loss": 3.704925537109375, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.01925475336611271, + "learning_rate": 0.00046393713164710217, + "loss": 3.711491346359253, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019796503707766533, + "learning_rate": 0.00046371301188168204, + "loss": 3.7241992950439453, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.018511971458792686, + "learning_rate": 0.0004634887626326911, + "loss": 3.6785435676574707, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.018214019015431404, + "learning_rate": 0.00046326438408080293, + "loss": 3.7299962043762207, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.018912190571427345, + "learning_rate": 0.00046303987640679517, + "loss": 3.773698329925537, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.01939854398369789, + "learning_rate": 0.0004628152397915498, + "loss": 3.682321071624756, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.017859861254692078, + "learning_rate": 0.00046259047441605215, + "loss": 3.7254951000213623, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.019249707460403442, + "learning_rate": 0.0004623655804613919, + "loss": 3.7263433933258057, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.019221967086195946, + "learning_rate": 0.00046214055810876194, + "loss": 3.7229225635528564, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01871953159570694, + "learning_rate": 0.00046191540753945886, + "loss": 3.7248311042785645, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01747182011604309, + "learning_rate": 0.00046169012893488204, + "loss": 3.699913740158081, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.019630730152130127, + "learning_rate": 0.00046146472247653484, + "loss": 3.787783145904541, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.072091102600098, + "eval_runtime": 63.1589, + "eval_samples_per_second": 38.664, + "eval_steps_per_second": 1.219, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.022918934002518654, + "learning_rate": 0.0004612391883460228, + "loss": 3.666675567626953, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.0264838095754385, + "learning_rate": 0.00046101352672505493, + "loss": 3.6371259689331055, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.028463594615459442, + "learning_rate": 0.0004607877377954426, + "loss": 3.657130241394043, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.02900792844593525, + "learning_rate": 0.00046056182173909995, + "loss": 3.6365628242492676, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.028593607246875763, + "learning_rate": 0.0004603357787380435, + "loss": 3.647258996963501, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.02878790907561779, + "learning_rate": 0.0004601096089743919, + "loss": 3.682035446166992, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.02980183996260166, + "learning_rate": 0.0004598833126303661, + "loss": 3.623297691345215, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.03375725448131561, + "learning_rate": 0.00045965688988828884, + "loss": 3.6240830421447754, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.035045325756073, + "learning_rate": 0.00045943034093058507, + "loss": 3.6372249126434326, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.03519923985004425, + "learning_rate": 0.000459203665939781, + "loss": 3.6157641410827637, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.03468473628163338, + "learning_rate": 0.00045897686509850487, + "loss": 3.6553335189819336, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.034817393869161606, + "learning_rate": 0.0004587499385894856, + "loss": 3.653806686401367, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.03549059107899666, + "learning_rate": 0.000458522886595554, + "loss": 3.6730189323425293, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.028001410886645317, + "learning_rate": 0.0004582957092996418, + "loss": 3.662490129470825, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.02532520331442356, + "learning_rate": 0.0004580684068847817, + "loss": 3.629936456680298, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.028215918689966202, + "learning_rate": 0.0004578409795341069, + "loss": 3.6507444381713867, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.026823483407497406, + "learning_rate": 0.0004576134274308517, + "loss": 3.661811113357544, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.0281058382242918, + "learning_rate": 0.0004573857507583506, + "loss": 3.6805872917175293, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.02548503689467907, + "learning_rate": 0.00045715794970003874, + "loss": 3.6586413383483887, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.024795135483145714, + "learning_rate": 0.0004569300244394513, + "loss": 3.654989004135132, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.025501688942313194, + "learning_rate": 0.0004567019751602234, + "loss": 3.6915903091430664, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.02377370186150074, + "learning_rate": 0.00045647380204609034, + "loss": 3.660588264465332, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.022150425240397453, + "learning_rate": 0.00045624550528088704, + "loss": 3.64870023727417, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.02415974996984005, + "learning_rate": 0.00045601708504854803, + "loss": 3.6498191356658936, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.02140078693628311, + "learning_rate": 0.0004557885415331074, + "loss": 3.703489065170288, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.022121304646134377, + "learning_rate": 0.00045555987491869865, + "loss": 3.6606979370117188, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.02168990671634674, + "learning_rate": 0.00045533108538955416, + "loss": 3.6801178455352783, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.020228713750839233, + "learning_rate": 0.0004551021731300055, + "loss": 3.6707205772399902, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.018500352278351784, + "learning_rate": 0.0004548731383244834, + "loss": 3.647545576095581, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.020223816856741905, + "learning_rate": 0.0004546439811575169, + "loss": 3.644932985305786, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.019502121955156326, + "learning_rate": 0.00045441470181373375, + "loss": 3.62888240814209, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.01926770992577076, + "learning_rate": 0.0004541853004778603, + "loss": 3.675457715988159, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01888471283018589, + "learning_rate": 0.00045395577733472087, + "loss": 3.636434316635132, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.0201494712382555, + "learning_rate": 0.0004537261325692383, + "loss": 3.679147720336914, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.02178201824426651, + "learning_rate": 0.0004534963663664332, + "loss": 3.671905040740967, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.024073559790849686, + "learning_rate": 0.00045326647891142394, + "loss": 3.6701040267944336, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.02135559171438217, + "learning_rate": 0.0004530364703894268, + "loss": 3.648519992828369, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.019264448434114456, + "learning_rate": 0.0004528063409857554, + "loss": 3.6985023021698, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.01922295056283474, + "learning_rate": 0.0004525760908858209, + "loss": 3.651756763458252, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.01978331431746483, + "learning_rate": 0.00045234572027513174, + "loss": 3.679884910583496, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.01968599483370781, + "learning_rate": 0.00045211522933929317, + "loss": 3.6571812629699707, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.020741639658808708, + "learning_rate": 0.0004518846182640077, + "loss": 3.6505298614501953, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.01842017099261284, + "learning_rate": 0.0004516538872350744, + "loss": 3.6904940605163574, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.018004486337304115, + "learning_rate": 0.0004514230364383893, + "loss": 3.7282047271728516, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.018821008503437042, + "learning_rate": 0.0004511920660599444, + "loss": 3.668164014816284, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.018929854035377502, + "learning_rate": 0.0004509609762858287, + "loss": 3.6672980785369873, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.01979168877005577, + "learning_rate": 0.0004507297673022269, + "loss": 3.6684999465942383, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.01774384267628193, + "learning_rate": 0.0004504984392954199, + "loss": 3.6577749252319336, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01876000687479973, + "learning_rate": 0.00045026699245178444, + "loss": 3.6629252433776855, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.018197249621152878, + "learning_rate": 0.0004500354269577932, + "loss": 3.688983678817749, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.019496265798807144, + "learning_rate": 0.00044980374300001414, + "loss": 3.6602022647857666, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01825535111129284, + "learning_rate": 0.00044957194076511095, + "loss": 3.6889238357543945, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.017113137990236282, + "learning_rate": 0.0004493400204398426, + "loss": 3.684070587158203, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.017993126064538956, + "learning_rate": 0.00044910798221106283, + "loss": 3.675426721572876, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.018354708328843117, + "learning_rate": 0.00044887582626572105, + "loss": 3.675588607788086, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.016362832859158516, + "learning_rate": 0.0004486435527908606, + "loss": 3.6709303855895996, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01772182621061802, + "learning_rate": 0.0004484111619736205, + "loss": 3.6884102821350098, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.019271764904260635, + "learning_rate": 0.0004481786540012336, + "loss": 3.6468968391418457, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.019040528684854507, + "learning_rate": 0.00044794602906102747, + "loss": 3.690387010574341, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.019022610038518906, + "learning_rate": 0.00044771328734042367, + "loss": 3.689873218536377, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.018533391878008842, + "learning_rate": 0.00044748042902693815, + "loss": 3.6877284049987793, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.020825615152716637, + "learning_rate": 0.0004472474543081805, + "loss": 3.662257671356201, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.021880341693758965, + "learning_rate": 0.00044701436337185444, + "loss": 3.650700092315674, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.019974704831838608, + "learning_rate": 0.00044678115640575704, + "loss": 3.684177875518799, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.01903127320110798, + "learning_rate": 0.0004465478335977788, + "loss": 3.658148765563965, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.020999105647206306, + "learning_rate": 0.00044631439513590393, + "loss": 3.68747615814209, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.02138398587703705, + "learning_rate": 0.00044608084120820933, + "loss": 3.656355381011963, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.02216442860662937, + "learning_rate": 0.0004458471720028655, + "loss": 3.6860342025756836, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.019359121099114418, + "learning_rate": 0.0004456133877081352, + "loss": 3.654944658279419, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.020984770730137825, + "learning_rate": 0.0004453794885123744, + "loss": 3.6885194778442383, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.02423972077667713, + "learning_rate": 0.00044514547460403135, + "loss": 3.672712802886963, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.024544453248381615, + "learning_rate": 0.00044491134617164694, + "loss": 3.6484768390655518, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.060352325439453, + "eval_runtime": 57.9556, + "eval_samples_per_second": 42.136, + "eval_steps_per_second": 1.329, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.024072550237178802, + "learning_rate": 0.000444677103403854, + "loss": 3.5796468257904053, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.028701433911919594, + "learning_rate": 0.0004444427464893779, + "loss": 3.604318618774414, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.028119578957557678, + "learning_rate": 0.00044420827561703573, + "loss": 3.608759641647339, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.025335192680358887, + "learning_rate": 0.0004439736909757365, + "loss": 3.5760319232940674, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.0273929201066494, + "learning_rate": 0.0004437389927544805, + "loss": 3.559441089630127, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.026567714288830757, + "learning_rate": 0.0004435041811423603, + "loss": 3.626889228820801, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.027587885037064552, + "learning_rate": 0.0004432692563285591, + "loss": 3.6226272583007812, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.02935059368610382, + "learning_rate": 0.0004430342185023517, + "loss": 3.601921558380127, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.02982175536453724, + "learning_rate": 0.0004427990678531038, + "loss": 3.59737491607666, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.028570258989930153, + "learning_rate": 0.0004425638045702719, + "loss": 3.56337833404541, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.029890350997447968, + "learning_rate": 0.0004423284288434035, + "loss": 3.6294307708740234, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.03206814453005791, + "learning_rate": 0.00044209294086213657, + "loss": 3.590505599975586, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.03169345110654831, + "learning_rate": 0.00044185734081619957, + "loss": 3.581584930419922, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.03313758969306946, + "learning_rate": 0.000441621628895411, + "loss": 3.6088438034057617, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.03317603841423988, + "learning_rate": 0.00044138580528967986, + "loss": 3.6340456008911133, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.034738484770059586, + "learning_rate": 0.0004411498701890051, + "loss": 3.5814096927642822, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.036428093910217285, + "learning_rate": 0.0004409138237834751, + "loss": 3.5939526557922363, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.03293406963348389, + "learning_rate": 0.00044067766626326835, + "loss": 3.577897787094116, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.031030720099806786, + "learning_rate": 0.00044044139781865255, + "loss": 3.6077990531921387, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.03441236540675163, + "learning_rate": 0.00044020501863998516, + "loss": 3.591923713684082, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.03500965237617493, + "learning_rate": 0.00043996852891771236, + "loss": 3.6171927452087402, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.034443050622940063, + "learning_rate": 0.00043973192884236983, + "loss": 3.619724750518799, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.03435051813721657, + "learning_rate": 0.0004394952186045819, + "loss": 3.6443169116973877, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.03230702877044678, + "learning_rate": 0.00043925839839506167, + "loss": 3.5912671089172363, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.03232988342642784, + "learning_rate": 0.0004390214684046111, + "loss": 3.618985414505005, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.029298555105924606, + "learning_rate": 0.00043878442882412015, + "loss": 3.6166677474975586, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.026723450049757957, + "learning_rate": 0.00043854727984456746, + "loss": 3.61362886428833, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.028280874714255333, + "learning_rate": 0.0004383100216570197, + "loss": 3.6410951614379883, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.026700228452682495, + "learning_rate": 0.0004380726544526316, + "loss": 3.5804049968719482, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.023092713207006454, + "learning_rate": 0.00043783517842264534, + "loss": 3.595339298248291, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.024123897776007652, + "learning_rate": 0.0004375975937583914, + "loss": 3.647569417953491, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.02505572699010372, + "learning_rate": 0.0004373599006512871, + "loss": 3.6276652812957764, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.024523509666323662, + "learning_rate": 0.00043712209929283786, + "loss": 3.59525728225708, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.024351581931114197, + "learning_rate": 0.00043688418987463567, + "loss": 3.6081674098968506, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.02530614472925663, + "learning_rate": 0.00043664617258835995, + "loss": 3.6632542610168457, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.023438062518835068, + "learning_rate": 0.0004364080476257769, + "loss": 3.6345431804656982, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.021853657439351082, + "learning_rate": 0.0004361698151787396, + "loss": 3.598644256591797, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.022487422451376915, + "learning_rate": 0.0004359314754391874, + "loss": 3.5966577529907227, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.021353935822844505, + "learning_rate": 0.0004356930285991466, + "loss": 3.6328375339508057, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.021180452778935432, + "learning_rate": 0.00043545447485072926, + "loss": 3.6285812854766846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.020491832867264748, + "learning_rate": 0.000435215814386134, + "loss": 3.592944622039795, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.021603770554065704, + "learning_rate": 0.0004349770473976453, + "loss": 3.613507032394409, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.02146158367395401, + "learning_rate": 0.0004347381740776332, + "loss": 3.6409506797790527, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.02102433145046234, + "learning_rate": 0.0004344991946185538, + "loss": 3.6179330348968506, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.018684551119804382, + "learning_rate": 0.00043426010921294837, + "loss": 3.6385130882263184, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.018416227772831917, + "learning_rate": 0.0004340209180534438, + "loss": 3.642151355743408, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.01915144734084606, + "learning_rate": 0.000433781621332752, + "loss": 3.6089656352996826, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.018299665302038193, + "learning_rate": 0.0004335422192436703, + "loss": 3.6063547134399414, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.018962757661938667, + "learning_rate": 0.00043330271197908023, + "loss": 3.625917434692383, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.017437821254134178, + "learning_rate": 0.00043306309973194874, + "loss": 3.600015640258789, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.017913147807121277, + "learning_rate": 0.000432823382695327, + "loss": 3.6428744792938232, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.018343618139624596, + "learning_rate": 0.0004325835610623508, + "loss": 3.635901927947998, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.020372850820422173, + "learning_rate": 0.00043234363502623994, + "loss": 3.620181083679199, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.02037600241601467, + "learning_rate": 0.0004321036047802985, + "loss": 3.61129093170166, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.021827151998877525, + "learning_rate": 0.00043186347051791466, + "loss": 3.5999555587768555, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.021806657314300537, + "learning_rate": 0.0004316232324325602, + "loss": 3.642509937286377, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02050699107348919, + "learning_rate": 0.0004313828907177906, + "loss": 3.597583293914795, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.020436758175492287, + "learning_rate": 0.0004311424455672448, + "loss": 3.6079912185668945, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.019341666251420975, + "learning_rate": 0.0004309018971746451, + "loss": 3.6196017265319824, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.022348817437887192, + "learning_rate": 0.000430661245733797, + "loss": 3.619025468826294, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.019361121580004692, + "learning_rate": 0.00043042049143858914, + "loss": 3.657182455062866, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.019258935004472733, + "learning_rate": 0.0004301796344829928, + "loss": 3.6408963203430176, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.020622272044420242, + "learning_rate": 0.00042993867506106205, + "loss": 3.6247220039367676, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.019510755315423012, + "learning_rate": 0.0004296976133669336, + "loss": 3.6264584064483643, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.02078665979206562, + "learning_rate": 0.0004294564495948266, + "loss": 3.6364388465881348, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01791800744831562, + "learning_rate": 0.0004292151839390421, + "loss": 3.6029374599456787, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.01922667771577835, + "learning_rate": 0.0004289738165939636, + "loss": 3.582937717437744, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.018473893404006958, + "learning_rate": 0.00042873234775405654, + "loss": 3.6141819953918457, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01846298575401306, + "learning_rate": 0.00042849077761386784, + "loss": 3.638786554336548, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.01876973733305931, + "learning_rate": 0.0004282491063680263, + "loss": 3.617130756378174, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.017995499074459076, + "learning_rate": 0.00042800733421124207, + "loss": 3.645305633544922, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.02105429396033287, + "learning_rate": 0.00042776546133830646, + "loss": 3.6444876194000244, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.049062252044678, + "eval_runtime": 58.0113, + "eval_samples_per_second": 42.095, + "eval_steps_per_second": 1.327, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.02128119394183159, + "learning_rate": 0.00042752348794409227, + "loss": 3.524954080581665, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.028249500319361687, + "learning_rate": 0.00042728141422355287, + "loss": 3.5311479568481445, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.03145391866564751, + "learning_rate": 0.0004270392403717229, + "loss": 3.559330940246582, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.02994118444621563, + "learning_rate": 0.00042679696658371743, + "loss": 3.5257620811462402, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.027448246255517006, + "learning_rate": 0.00042655459305473195, + "loss": 3.5242764949798584, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.0296134352684021, + "learning_rate": 0.0004263121199800425, + "loss": 3.5317091941833496, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.029576804488897324, + "learning_rate": 0.0004260695475550054, + "loss": 3.5340428352355957, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.02941269613802433, + "learning_rate": 0.0004258268759750566, + "loss": 3.5372023582458496, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.03126693144440651, + "learning_rate": 0.0004255841054357124, + "loss": 3.5437135696411133, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.03335367515683174, + "learning_rate": 0.0004253412361325686, + "loss": 3.5598273277282715, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.03028946928679943, + "learning_rate": 0.0004250982682613006, + "loss": 3.5190839767456055, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.028629133477807045, + "learning_rate": 0.00042485520201766315, + "loss": 3.5138821601867676, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.027070026844739914, + "learning_rate": 0.0004246120375974905, + "loss": 3.530958414077759, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.028837760910391808, + "learning_rate": 0.0004243687751966957, + "loss": 3.5233962535858154, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.029884297400712967, + "learning_rate": 0.0004241254150112708, + "loss": 3.5639758110046387, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.024223247542977333, + "learning_rate": 0.000423881957237287, + "loss": 3.5555152893066406, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.026992836967110634, + "learning_rate": 0.0004236384020708935, + "loss": 3.540196418762207, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.02818330004811287, + "learning_rate": 0.0004233947497083185, + "loss": 3.5546071529388428, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.026401042938232422, + "learning_rate": 0.0004231510003458682, + "loss": 3.5556247234344482, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.026020489633083344, + "learning_rate": 0.0004229071541799272, + "loss": 3.568004608154297, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.02803782932460308, + "learning_rate": 0.0004226632114069578, + "loss": 3.5828235149383545, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.02737978659570217, + "learning_rate": 0.0004224191722235004, + "loss": 3.5642499923706055, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.02406371757388115, + "learning_rate": 0.00042217503682617286, + "loss": 3.5530993938446045, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.02230692468583584, + "learning_rate": 0.0004219308054116706, + "loss": 3.5573501586914062, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.021352406591176987, + "learning_rate": 0.0004216864781767667, + "loss": 3.5526974201202393, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.02227999083697796, + "learning_rate": 0.00042144205531831103, + "loss": 3.5866928100585938, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.024190247058868408, + "learning_rate": 0.0004211975370332308, + "loss": 3.517117977142334, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.023196619004011154, + "learning_rate": 0.00042095292351852976, + "loss": 3.4826502799987793, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.02388714998960495, + "learning_rate": 0.0004207082149712888, + "loss": 3.5588626861572266, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.024692339822649956, + "learning_rate": 0.0004204634115886651, + "loss": 3.54828143119812, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.023078206926584244, + "learning_rate": 0.0004202185135678924, + "loss": 3.5679030418395996, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.020665735006332397, + "learning_rate": 0.00041997352110628033, + "loss": 3.5500283241271973, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.02022322081029415, + "learning_rate": 0.00041972843440121527, + "loss": 3.592104911804199, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.021137656643986702, + "learning_rate": 0.00041948325365015894, + "loss": 3.583833694458008, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.021026931703090668, + "learning_rate": 0.0004192379790506491, + "loss": 3.556403875350952, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.020423648878932, + "learning_rate": 0.00041899261080029904, + "loss": 3.5839173793792725, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.020935002714395523, + "learning_rate": 0.00041874714909679754, + "loss": 3.5794217586517334, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.020799759775400162, + "learning_rate": 0.00041850159413790863, + "loss": 3.6033859252929688, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.019480884075164795, + "learning_rate": 0.0004182559461214715, + "loss": 3.580097198486328, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.019300777465105057, + "learning_rate": 0.0004180102052454004, + "loss": 3.550474166870117, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.020265772938728333, + "learning_rate": 0.00041776437170768427, + "loss": 3.5711655616760254, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.019972780719399452, + "learning_rate": 0.0004175184457063868, + "loss": 3.5636885166168213, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.019635623320937157, + "learning_rate": 0.000417272427439646, + "loss": 3.5851311683654785, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.021451901644468307, + "learning_rate": 0.00041702631710567443, + "loss": 3.5956759452819824, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.019456414505839348, + "learning_rate": 0.00041678011490275875, + "loss": 3.595731258392334, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.019757689908146858, + "learning_rate": 0.00041653382102925957, + "loss": 3.569652557373047, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.020571604371070862, + "learning_rate": 0.00041628743568361147, + "loss": 3.6007747650146484, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.02153848111629486, + "learning_rate": 0.00041604095906432265, + "loss": 3.5915260314941406, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.020875850692391396, + "learning_rate": 0.00041579439136997493, + "loss": 3.5582690238952637, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.02089419960975647, + "learning_rate": 0.00041554773279922337, + "loss": 3.5756618976593018, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01961098238825798, + "learning_rate": 0.0004153009835507963, + "loss": 3.5918776988983154, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.021075716242194176, + "learning_rate": 0.0004150541438234952, + "loss": 3.5342202186584473, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.020745476707816124, + "learning_rate": 0.00041480721381619435, + "loss": 3.5976028442382812, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.02001352794468403, + "learning_rate": 0.00041456019372784086, + "loss": 3.5686182975769043, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.021684160456061363, + "learning_rate": 0.0004143130837574543, + "loss": 3.553445339202881, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.022551920264959335, + "learning_rate": 0.0004140658841041267, + "loss": 3.5712709426879883, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.022758224979043007, + "learning_rate": 0.0004138185949670224, + "loss": 3.6101667881011963, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.020714636892080307, + "learning_rate": 0.00041357121654537766, + "loss": 3.5740506649017334, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.022694187238812447, + "learning_rate": 0.00041332374903850104, + "loss": 3.5663979053497314, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.021664340049028397, + "learning_rate": 0.00041307619264577234, + "loss": 3.5830888748168945, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.019072270020842552, + "learning_rate": 0.0004128285475666436, + "loss": 3.5861973762512207, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.020153764635324478, + "learning_rate": 0.0004125808140006378, + "loss": 3.566444158554077, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019666558131575584, + "learning_rate": 0.0004123329921473494, + "loss": 3.5789592266082764, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.02101624570786953, + "learning_rate": 0.0004120850822064439, + "loss": 3.551722526550293, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.01951020397245884, + "learning_rate": 0.00041183708437765815, + "loss": 3.6069014072418213, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.01985476352274418, + "learning_rate": 0.00041158899886079925, + "loss": 3.5944924354553223, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02195335365831852, + "learning_rate": 0.00041134082585574534, + "loss": 3.5828046798706055, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02105172537267208, + "learning_rate": 0.00041109256556244497, + "loss": 3.5830953121185303, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.01888827420771122, + "learning_rate": 0.00041084421818091685, + "loss": 3.5656847953796387, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.017867228016257286, + "learning_rate": 0.0004105957839112501, + "loss": 3.5792243480682373, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.018577905371785164, + "learning_rate": 0.0004103472629536036, + "loss": 3.56992506980896, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.020045632496476173, + "learning_rate": 0.0004100986555082063, + "loss": 3.590876579284668, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.044339656829834, + "eval_runtime": 57.1435, + "eval_samples_per_second": 42.735, + "eval_steps_per_second": 1.347, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.024135645478963852, + "learning_rate": 0.0004098499617753567, + "loss": 3.4856467247009277, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.031053155660629272, + "learning_rate": 0.00040960118195542273, + "loss": 3.455773115158081, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.035190410912036896, + "learning_rate": 0.00040935231624884204, + "loss": 3.506619930267334, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.03611094877123833, + "learning_rate": 0.0004091033648561211, + "loss": 3.5227696895599365, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.034121073782444, + "learning_rate": 0.00040885432797783543, + "loss": 3.4970145225524902, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.03629382699728012, + "learning_rate": 0.0004086052058146297, + "loss": 3.506425619125366, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.03709045797586441, + "learning_rate": 0.00040835599856721725, + "loss": 3.499274969100952, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.034826792776584625, + "learning_rate": 0.00040810670643637954, + "loss": 3.5029163360595703, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.03344690054655075, + "learning_rate": 0.000407857329622967, + "loss": 3.4816911220550537, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.03136949986219406, + "learning_rate": 0.00040760786832789786, + "loss": 3.5154659748077393, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.030533097684383392, + "learning_rate": 0.0004073583227521588, + "loss": 3.509666919708252, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.03298318013548851, + "learning_rate": 0.0004071086930968039, + "loss": 3.5096616744995117, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.031186232343316078, + "learning_rate": 0.00040685897956295545, + "loss": 3.4752590656280518, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.028008371591567993, + "learning_rate": 0.0004066091823518031, + "loss": 3.498826503753662, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.028206519782543182, + "learning_rate": 0.00040635930166460385, + "loss": 3.5281882286071777, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.024417119100689888, + "learning_rate": 0.00040610933770268226, + "loss": 3.506791830062866, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.025321070104837418, + "learning_rate": 0.00040585929066742964, + "loss": 3.494256019592285, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.023430505767464638, + "learning_rate": 0.00040560916076030435, + "loss": 3.533010482788086, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.023371374234557152, + "learning_rate": 0.00040535894818283154, + "loss": 3.5197086334228516, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.02414027415215969, + "learning_rate": 0.0004051086531366031, + "loss": 3.4707889556884766, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.026652060449123383, + "learning_rate": 0.000404858275823277, + "loss": 3.498159885406494, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.024486590176820755, + "learning_rate": 0.000404607816444578, + "loss": 3.50034236907959, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.023153992369771004, + "learning_rate": 0.00040435727520229644, + "loss": 3.499504804611206, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.02222427912056446, + "learning_rate": 0.0004041066522982891, + "loss": 3.4953579902648926, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.023033641278743744, + "learning_rate": 0.000403855947934478, + "loss": 3.5181496143341064, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.02141181193292141, + "learning_rate": 0.00040360516231285155, + "loss": 3.552860736846924, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.023448696359992027, + "learning_rate": 0.00040335429563546286, + "loss": 3.5153324604034424, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.022410666570067406, + "learning_rate": 0.0004031033481044308, + "loss": 3.4963512420654297, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.020326802507042885, + "learning_rate": 0.00040285231992193924, + "loss": 3.528319835662842, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021117394790053368, + "learning_rate": 0.0004026012112902372, + "loss": 3.5192837715148926, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.021317072212696075, + "learning_rate": 0.0004023500224116381, + "loss": 3.5081021785736084, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.021116144955158234, + "learning_rate": 0.00040209875348852037, + "loss": 3.5167174339294434, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019834019243717194, + "learning_rate": 0.00040184740472332705, + "loss": 3.511960983276367, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.02089378982782364, + "learning_rate": 0.000401595976318565, + "loss": 3.493330955505371, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.021129710599780083, + "learning_rate": 0.00040134446847680584, + "loss": 3.544156551361084, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.021163299679756165, + "learning_rate": 0.0004010928814006846, + "loss": 3.5261003971099854, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.02105306088924408, + "learning_rate": 0.0004008412152929007, + "loss": 3.543973445892334, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.02150859497487545, + "learning_rate": 0.0004005894703562168, + "loss": 3.487217426300049, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.021666452288627625, + "learning_rate": 0.0004003376467934593, + "loss": 3.4933671951293945, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.02180723287165165, + "learning_rate": 0.0004000857448075178, + "loss": 3.55198073387146, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.02041625790297985, + "learning_rate": 0.0003998337646013455, + "loss": 3.529571533203125, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.022977063432335854, + "learning_rate": 0.00039958170637795806, + "loss": 3.526432752609253, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.022725161164999008, + "learning_rate": 0.0003993295703404341, + "loss": 3.5258829593658447, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.021337240934371948, + "learning_rate": 0.00039907735669191544, + "loss": 3.4897351264953613, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.020783131942152977, + "learning_rate": 0.00039882506563560573, + "loss": 3.5497589111328125, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.022259317338466644, + "learning_rate": 0.0003985726973747715, + "loss": 3.5225064754486084, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.020283272489905357, + "learning_rate": 0.00039832025211274113, + "loss": 3.5453078746795654, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.020656265318393707, + "learning_rate": 0.00039806773005290544, + "loss": 3.533703327178955, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.02104976586997509, + "learning_rate": 0.0003978151313987168, + "loss": 3.4885146617889404, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.021411150693893433, + "learning_rate": 0.00039756245635368933, + "loss": 3.533268690109253, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.021906673908233643, + "learning_rate": 0.00039730970512139876, + "loss": 3.5483224391937256, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.02415728010237217, + "learning_rate": 0.0003970568779054822, + "loss": 3.5084290504455566, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.023435061797499657, + "learning_rate": 0.000396803974909638, + "loss": 3.5498428344726562, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.02183385379612446, + "learning_rate": 0.0003965509963376255, + "loss": 3.484086751937866, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.022856134921312332, + "learning_rate": 0.000396297942393265, + "loss": 3.5339231491088867, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.021679896861314774, + "learning_rate": 0.0003960448132804375, + "loss": 3.547191619873047, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.021262124180793762, + "learning_rate": 0.0003957916092030845, + "loss": 3.5263822078704834, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.02265402302145958, + "learning_rate": 0.00039553833036520803, + "loss": 3.5361135005950928, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.020080721005797386, + "learning_rate": 0.0003952849769708702, + "loss": 3.5247459411621094, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.02187187783420086, + "learning_rate": 0.0003950315492241932, + "loss": 3.54129695892334, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.02417500875890255, + "learning_rate": 0.0003947780473293593, + "loss": 3.598796844482422, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.02416270412504673, + "learning_rate": 0.00039452447149061054, + "loss": 3.5207936763763428, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.021958643570542336, + "learning_rate": 0.0003942708219122481, + "loss": 3.5395452976226807, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.021813420578837395, + "learning_rate": 0.000394017098798633, + "loss": 3.5469255447387695, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020607495680451393, + "learning_rate": 0.0003937633023541854, + "loss": 3.554851531982422, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.02091527171432972, + "learning_rate": 0.0003935094327833845, + "loss": 3.492576837539673, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.021089257672429085, + "learning_rate": 0.0003932554902907683, + "loss": 3.5154709815979004, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.019659532234072685, + "learning_rate": 0.0003930014750809338, + "loss": 3.5444283485412598, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01896163448691368, + "learning_rate": 0.0003927473873585365, + "loss": 3.589625358581543, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020288823172450066, + "learning_rate": 0.0003924932273282903, + "loss": 3.522411823272705, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.020222362130880356, + "learning_rate": 0.00039223899519496723, + "loss": 3.5244359970092773, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.022349944338202477, + "learning_rate": 0.00039198469116339757, + "loss": 3.5362396240234375, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.047028064727783, + "eval_runtime": 62.8948, + "eval_samples_per_second": 38.827, + "eval_steps_per_second": 1.224, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.02202913723886013, + "learning_rate": 0.0003917303154384694, + "loss": 3.4249138832092285, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.03153969347476959, + "learning_rate": 0.00039147586822512885, + "loss": 3.4436073303222656, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.03747870773077011, + "learning_rate": 0.0003912213497283793, + "loss": 3.4879579544067383, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.036144088953733444, + "learning_rate": 0.0003909667601532819, + "loss": 3.419151782989502, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.03983336314558983, + "learning_rate": 0.00039071209970495445, + "loss": 3.4331436157226562, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.045912645757198334, + "learning_rate": 0.00039045736858857273, + "loss": 3.4304769039154053, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.04005753993988037, + "learning_rate": 0.0003902025670093687, + "loss": 3.478600025177002, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.03488151356577873, + "learning_rate": 0.00038994769517263156, + "loss": 3.4650232791900635, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.03603382781147957, + "learning_rate": 0.0003896927532837069, + "loss": 3.449807643890381, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.03817906603217125, + "learning_rate": 0.0003894377415479967, + "loss": 3.4556589126586914, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.03523671627044678, + "learning_rate": 0.00038918266017095936, + "loss": 3.477426767349243, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.03284068778157234, + "learning_rate": 0.0003889275093581096, + "loss": 3.478672981262207, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.02992352657020092, + "learning_rate": 0.00038867228931501737, + "loss": 3.4810047149658203, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.02858242578804493, + "learning_rate": 0.0003884170002473091, + "loss": 3.467811107635498, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.027912171557545662, + "learning_rate": 0.0003881616423606666, + "loss": 3.4326171875, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.028351375833153725, + "learning_rate": 0.00038790621586082697, + "loss": 3.4533729553222656, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.027947887778282166, + "learning_rate": 0.0003876507209535829, + "loss": 3.420295476913452, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.028510035946965218, + "learning_rate": 0.00038739515784478186, + "loss": 3.438023328781128, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.027707934379577637, + "learning_rate": 0.00038713952674032654, + "loss": 3.4844181537628174, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.0262764822691679, + "learning_rate": 0.0003868838278461743, + "loss": 3.4363749027252197, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.02713681012392044, + "learning_rate": 0.0003866280613683371, + "loss": 3.4501900672912598, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.0299485195428133, + "learning_rate": 0.00038637222751288126, + "loss": 3.461942672729492, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.027047986164689064, + "learning_rate": 0.0003861163264859277, + "loss": 3.454293727874756, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.02436736971139908, + "learning_rate": 0.0003858603584936511, + "loss": 3.4663405418395996, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.02601189725100994, + "learning_rate": 0.0003856043237422803, + "loss": 3.502030849456787, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.02463468722999096, + "learning_rate": 0.0003853482224380978, + "loss": 3.462491035461426, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.021793190389871597, + "learning_rate": 0.00038509205478744, + "loss": 3.4476377964019775, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.023809107020497322, + "learning_rate": 0.0003848358209966963, + "loss": 3.4701178073883057, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.02557182125747204, + "learning_rate": 0.0003845795212723099, + "loss": 3.4932539463043213, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.023424552753567696, + "learning_rate": 0.00038432315582077664, + "loss": 3.4779105186462402, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.02195148915052414, + "learning_rate": 0.0003840667248486456, + "loss": 3.470506191253662, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.022897250950336456, + "learning_rate": 0.0003838102285625186, + "loss": 3.48982572555542, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.023033805191516876, + "learning_rate": 0.00038355366716905006, + "loss": 3.4845101833343506, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.022774485871195793, + "learning_rate": 0.0003832970408749467, + "loss": 3.5060324668884277, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.02193611115217209, + "learning_rate": 0.00038304034988696816, + "loss": 3.511518955230713, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.02130570448935032, + "learning_rate": 0.00038278359441192516, + "loss": 3.468502998352051, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.020402755588293076, + "learning_rate": 0.00038252677465668136, + "loss": 3.4786908626556396, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.022664198651909828, + "learning_rate": 0.00038226989082815156, + "loss": 3.479928970336914, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.02454056777060032, + "learning_rate": 0.00038201294313330245, + "loss": 3.479947566986084, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.022096732631325722, + "learning_rate": 0.00038175593177915227, + "loss": 3.49478816986084, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.021499406546354294, + "learning_rate": 0.0003814988569727704, + "loss": 3.494933605194092, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.022325703874230385, + "learning_rate": 0.0003812417189212773, + "loss": 3.5041463375091553, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.020674455910921097, + "learning_rate": 0.00038098451783184447, + "loss": 3.472404718399048, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.02131054364144802, + "learning_rate": 0.00038072725391169416, + "loss": 3.497795343399048, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.02126218006014824, + "learning_rate": 0.0003804699273680994, + "loss": 3.483281373977661, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.0217830128967762, + "learning_rate": 0.0003802125384083834, + "loss": 3.4704298973083496, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.02156917378306389, + "learning_rate": 0.0003799550872399197, + "loss": 3.4994399547576904, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.02047055773437023, + "learning_rate": 0.0003796975740701323, + "loss": 3.4316084384918213, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.019766805693507195, + "learning_rate": 0.0003794399991064948, + "loss": 3.452223777770996, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.020028885453939438, + "learning_rate": 0.00037918236255653074, + "loss": 3.485278844833374, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.020312055945396423, + "learning_rate": 0.00037892466462781306, + "loss": 3.4789156913757324, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.020725280046463013, + "learning_rate": 0.00037866690552796446, + "loss": 3.495528221130371, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.02088077738881111, + "learning_rate": 0.00037840908546465696, + "loss": 3.4972569942474365, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.021010037511587143, + "learning_rate": 0.0003781512046456111, + "loss": 3.473562240600586, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.019641095772385597, + "learning_rate": 0.0003778932632785972, + "loss": 3.4782915115356445, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.02063329517841339, + "learning_rate": 0.00037763526157143376, + "loss": 3.4681782722473145, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.0213928259909153, + "learning_rate": 0.00037737719973198825, + "loss": 3.472263813018799, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.02135022170841694, + "learning_rate": 0.0003771190779681762, + "loss": 3.502300262451172, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.02053583413362503, + "learning_rate": 0.0003768608964879619, + "loss": 3.489243745803833, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.021575238555669785, + "learning_rate": 0.00037660265549935724, + "loss": 3.491074562072754, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.022335771471261978, + "learning_rate": 0.00037634435521042257, + "loss": 3.5298843383789062, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.021198464557528496, + "learning_rate": 0.0003760859958292656, + "loss": 3.5156898498535156, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.020986037328839302, + "learning_rate": 0.000375827577564042, + "loss": 3.4574475288391113, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.022639811038970947, + "learning_rate": 0.0003755691006229545, + "loss": 3.4952611923217773, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.019711066037416458, + "learning_rate": 0.0003753105652142534, + "loss": 3.531296730041504, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.02058061584830284, + "learning_rate": 0.000375051971546236, + "loss": 3.5340657234191895, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.019849255681037903, + "learning_rate": 0.0003747933198272465, + "loss": 3.5028295516967773, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01954430714249611, + "learning_rate": 0.00037453461026567604, + "loss": 3.51088809967041, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.019234873354434967, + "learning_rate": 0.00037427584306996196, + "loss": 3.472560167312622, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.021085912361741066, + "learning_rate": 0.0003740170184485888, + "loss": 3.468876838684082, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.020300406962633133, + "learning_rate": 0.0003737581366100864, + "loss": 3.4903886318206787, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02380351349711418, + "learning_rate": 0.0003734991977630315, + "loss": 3.5069704055786133, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.048111915588379, + "eval_runtime": 58.312, + "eval_samples_per_second": 41.878, + "eval_steps_per_second": 1.32, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02379327267408371, + "learning_rate": 0.0003732402021160463, + "loss": 3.396376132965088, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.03204222396016121, + "learning_rate": 0.00037298114987779885, + "loss": 3.4102582931518555, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.034993767738342285, + "learning_rate": 0.000372722041257003, + "loss": 3.385697364807129, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.03387003391981125, + "learning_rate": 0.00037246287646241783, + "loss": 3.4120078086853027, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.03282478079199791, + "learning_rate": 0.0003722036557028478, + "loss": 3.404122829437256, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.03053346462547779, + "learning_rate": 0.0003719443791871422, + "loss": 3.368051767349243, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.029722390696406364, + "learning_rate": 0.0003716850471241958, + "loss": 3.4066247940063477, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.0287226140499115, + "learning_rate": 0.0003714256597229474, + "loss": 3.4212646484375, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.028156550601124763, + "learning_rate": 0.0003711662171923809, + "loss": 3.4192593097686768, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.03021939843893051, + "learning_rate": 0.0003709067197415244, + "loss": 3.3973188400268555, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.03314025700092316, + "learning_rate": 0.00037064716757945036, + "loss": 3.422978401184082, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.03156686946749687, + "learning_rate": 0.0003703875609152753, + "loss": 3.4052343368530273, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.03357478231191635, + "learning_rate": 0.0003701278999581595, + "loss": 3.383730411529541, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.03261081501841545, + "learning_rate": 0.0003698681849173073, + "loss": 3.421250820159912, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.03252755105495453, + "learning_rate": 0.0003696084160019662, + "loss": 3.39884352684021, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.03177274018526077, + "learning_rate": 0.0003693485934214274, + "loss": 3.4175422191619873, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.028845487162470818, + "learning_rate": 0.0003690887173850253, + "loss": 3.3908143043518066, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.026045162230730057, + "learning_rate": 0.0003688287881021374, + "loss": 3.4417710304260254, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.027937524020671844, + "learning_rate": 0.00036856880578218376, + "loss": 3.410548210144043, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.025780271738767624, + "learning_rate": 0.00036830877063462786, + "loss": 3.4069018363952637, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.026093080639839172, + "learning_rate": 0.0003680486828689749, + "loss": 3.4199600219726562, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.025765180587768555, + "learning_rate": 0.00036778854269477315, + "loss": 3.442903995513916, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.022773101925849915, + "learning_rate": 0.0003675283503216127, + "loss": 3.394232749938965, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.025135623291134834, + "learning_rate": 0.000367268105959126, + "loss": 3.4466171264648438, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.02491890639066696, + "learning_rate": 0.00036700780981698705, + "loss": 3.418689727783203, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.02521841786801815, + "learning_rate": 0.0003667474621049119, + "loss": 3.4213061332702637, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.025186199694871902, + "learning_rate": 0.00036648706303265795, + "loss": 3.437631130218506, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.022097431123256683, + "learning_rate": 0.000366226612810024, + "loss": 3.4014902114868164, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.024009618908166885, + "learning_rate": 0.00036596611164685025, + "loss": 3.4292187690734863, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.022758696228265762, + "learning_rate": 0.0003657055597530175, + "loss": 3.4105381965637207, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.02447321265935898, + "learning_rate": 0.00036544495733844797, + "loss": 3.4196958541870117, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.025843804702162743, + "learning_rate": 0.000365184304613104, + "loss": 3.4343783855438232, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.022433951497077942, + "learning_rate": 0.0003649236017869892, + "loss": 3.440580368041992, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.02250702679157257, + "learning_rate": 0.00036466284907014685, + "loss": 3.4255261421203613, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.022953296080231667, + "learning_rate": 0.0003644020466726608, + "loss": 3.442258358001709, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.021977661177515984, + "learning_rate": 0.00036414119480465473, + "loss": 3.4058687686920166, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.021751178428530693, + "learning_rate": 0.0003638802936762925, + "loss": 3.4211649894714355, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.022764205932617188, + "learning_rate": 0.0003636193434977771, + "loss": 3.424694061279297, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.024827787652611732, + "learning_rate": 0.0003633583444793516, + "loss": 3.458076238632202, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.021888602524995804, + "learning_rate": 0.00036309729683129814, + "loss": 3.4509153366088867, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.021961702033877373, + "learning_rate": 0.00036283620076393814, + "loss": 3.4147253036499023, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.02235778421163559, + "learning_rate": 0.000362575056487632, + "loss": 3.4392640590667725, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.021714260801672935, + "learning_rate": 0.00036231386421277877, + "loss": 3.444026470184326, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.020996905863285065, + "learning_rate": 0.0003620526241498167, + "loss": 3.4168434143066406, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.020528919994831085, + "learning_rate": 0.0003617913365092218, + "loss": 3.4831130504608154, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.021056141704320908, + "learning_rate": 0.0003615300015015091, + "loss": 3.430905342102051, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.021655110642313957, + "learning_rate": 0.0003612686193372312, + "loss": 3.4588241577148438, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.02344612590968609, + "learning_rate": 0.00036100719022697924, + "loss": 3.4202680587768555, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.02264723740518093, + "learning_rate": 0.0003607457143813819, + "loss": 3.4669015407562256, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.02377854660153389, + "learning_rate": 0.00036048419201110553, + "loss": 3.5059216022491455, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.02299923449754715, + "learning_rate": 0.00036022262332685383, + "loss": 3.4796524047851562, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.021110547706484795, + "learning_rate": 0.0003599610085393681, + "loss": 3.4342408180236816, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.022060122340917587, + "learning_rate": 0.0003596993478594267, + "loss": 3.426923990249634, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.021077830344438553, + "learning_rate": 0.0003594376414978447, + "loss": 3.4636270999908447, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.02209082990884781, + "learning_rate": 0.0003591758896654745, + "loss": 3.441979169845581, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.022408071905374527, + "learning_rate": 0.0003589140925732045, + "loss": 3.490835666656494, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.022462666034698486, + "learning_rate": 0.0003586522504319602, + "loss": 3.452770233154297, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.021295800805091858, + "learning_rate": 0.0003583903634527029, + "loss": 3.4377408027648926, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.020457757636904716, + "learning_rate": 0.0003581284318464302, + "loss": 3.4191250801086426, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.021163633093237877, + "learning_rate": 0.00035786645582417564, + "loss": 3.4797091484069824, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.021495461463928223, + "learning_rate": 0.00035760443559700863, + "loss": 3.4763145446777344, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.021050071343779564, + "learning_rate": 0.00035734237137603417, + "loss": 3.4687209129333496, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.02150929905474186, + "learning_rate": 0.0003570802633723927, + "loss": 3.4819254875183105, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.0212967898696661, + "learning_rate": 0.0003568181117972597, + "loss": 3.490407705307007, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.02123691327869892, + "learning_rate": 0.00035655591686184626, + "loss": 3.4418535232543945, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.02193271741271019, + "learning_rate": 0.0003562936787773979, + "loss": 3.446892738342285, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.02065582573413849, + "learning_rate": 0.000356031397755195, + "loss": 3.4308691024780273, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.019487040117383003, + "learning_rate": 0.00035576907400655303, + "loss": 3.487239122390747, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.021190891042351723, + "learning_rate": 0.00035550670774282106, + "loss": 3.4635558128356934, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.0209872554987669, + "learning_rate": 0.0003552442991753831, + "loss": 3.443793773651123, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.02035645768046379, + "learning_rate": 0.00035498184851565697, + "loss": 3.473198652267456, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.024002201855182648, + "learning_rate": 0.0003547193559750944, + "loss": 3.4244751930236816, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.057085990905762, + "eval_runtime": 62.4735, + "eval_samples_per_second": 39.089, + "eval_steps_per_second": 1.233, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.02450242079794407, + "learning_rate": 0.00035445682176518087, + "loss": 3.361237049102783, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.030813416466116905, + "learning_rate": 0.00035419424609743546, + "loss": 3.365708351135254, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.03279578685760498, + "learning_rate": 0.00035393162918341065, + "loss": 3.325385570526123, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.03293970227241516, + "learning_rate": 0.00035366897123469213, + "loss": 3.356811285018921, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.033190157264471054, + "learning_rate": 0.0003534062724628986, + "loss": 3.406217098236084, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.033370476216077805, + "learning_rate": 0.00035314353307968195, + "loss": 3.3579213619232178, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.03351330757141113, + "learning_rate": 0.00035288075329672634, + "loss": 3.3755106925964355, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.034596871584653854, + "learning_rate": 0.0003526179333257487, + "loss": 3.3576536178588867, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.03177076578140259, + "learning_rate": 0.0003523550733784984, + "loss": 3.3123257160186768, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.0289607010781765, + "learning_rate": 0.00035209217366675706, + "loss": 3.370959997177124, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.03265267238020897, + "learning_rate": 0.000351829234402338, + "loss": 3.332608938217163, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.03397805988788605, + "learning_rate": 0.0003515662557970867, + "loss": 3.364213228225708, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.03100395016372204, + "learning_rate": 0.0003513032380628804, + "loss": 3.388005256652832, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.03224753588438034, + "learning_rate": 0.0003510401814116277, + "loss": 3.3267595767974854, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.03034576214849949, + "learning_rate": 0.0003507770860552684, + "loss": 3.3954129219055176, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.026316817849874496, + "learning_rate": 0.00035051395220577397, + "loss": 3.3543832302093506, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.027714375406503677, + "learning_rate": 0.0003502507800751464, + "loss": 3.4190099239349365, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.028731998056173325, + "learning_rate": 0.0003499875698754187, + "loss": 3.342031955718994, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02845810353755951, + "learning_rate": 0.00034972432181865467, + "loss": 3.356130599975586, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.02842753194272518, + "learning_rate": 0.00034946103611694854, + "loss": 3.3709869384765625, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.02942183054983616, + "learning_rate": 0.00034919771298242477, + "loss": 3.373588800430298, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.02958574891090393, + "learning_rate": 0.00034893435262723814, + "loss": 3.377514600753784, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.029524607583880424, + "learning_rate": 0.00034867095526357325, + "loss": 3.398663282394409, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.028092598542571068, + "learning_rate": 0.0003484075211036446, + "loss": 3.3742518424987793, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.026652829721570015, + "learning_rate": 0.0003481440503596964, + "loss": 3.387667179107666, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.024675756692886353, + "learning_rate": 0.0003478805432440021, + "loss": 3.3737070560455322, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.02491692826151848, + "learning_rate": 0.0003476169999688648, + "loss": 3.359217643737793, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.024601681157946587, + "learning_rate": 0.00034735342074661654, + "loss": 3.3902688026428223, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.02313913404941559, + "learning_rate": 0.0003470898057896183, + "loss": 3.3700339794158936, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.02418898604810238, + "learning_rate": 0.0003468261553102599, + "loss": 3.4265692234039307, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.023382030427455902, + "learning_rate": 0.00034656246952095984, + "loss": 3.378389835357666, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.02340470254421234, + "learning_rate": 0.0003462987486341648, + "loss": 3.392937421798706, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.022166339680552483, + "learning_rate": 0.00034603499286235006, + "loss": 3.39414119720459, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.022175129503011703, + "learning_rate": 0.0003457712024180188, + "loss": 3.407447099685669, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.02368207648396492, + "learning_rate": 0.0003455073775137025, + "loss": 3.415076971054077, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.02263982594013214, + "learning_rate": 0.0003452435183619598, + "loss": 3.383986473083496, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.02147712931036949, + "learning_rate": 0.0003449796251753773, + "loss": 3.403963565826416, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.022987166419625282, + "learning_rate": 0.00034471569816656915, + "loss": 3.3883113861083984, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.02240179106593132, + "learning_rate": 0.00034445173754817646, + "loss": 3.4349911212921143, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.02206592820584774, + "learning_rate": 0.00034418774353286747, + "loss": 3.426473617553711, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.022650036960840225, + "learning_rate": 0.0003439237163333375, + "loss": 3.4251773357391357, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.023399339988827705, + "learning_rate": 0.0003436596561623084, + "loss": 3.371610641479492, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.02192695252597332, + "learning_rate": 0.0003433955632325288, + "loss": 3.387253761291504, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.02241935208439827, + "learning_rate": 0.00034313143775677353, + "loss": 3.434523105621338, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.023202672600746155, + "learning_rate": 0.00034286727994784367, + "loss": 3.369771957397461, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.021333014592528343, + "learning_rate": 0.0003426030900185665, + "loss": 3.439004421234131, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.023988714441657066, + "learning_rate": 0.0003423388681817949, + "loss": 3.396838665008545, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.023885777220129967, + "learning_rate": 0.00034207461465040793, + "loss": 3.4114322662353516, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.02263713628053665, + "learning_rate": 0.0003418103296373096, + "loss": 3.4139351844787598, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.022404303774237633, + "learning_rate": 0.00034154601335542964, + "loss": 3.4200823307037354, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.021705076098442078, + "learning_rate": 0.00034128166601772304, + "loss": 3.4092609882354736, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.02273249812424183, + "learning_rate": 0.0003410172878371695, + "loss": 3.3931846618652344, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.022552402690052986, + "learning_rate": 0.00034075287902677394, + "loss": 3.396566867828369, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.023219924420118332, + "learning_rate": 0.0003404884397995655, + "loss": 3.369025707244873, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.022560900077223778, + "learning_rate": 0.00034022397036859837, + "loss": 3.4248225688934326, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.023370858281850815, + "learning_rate": 0.0003399594709469506, + "loss": 3.431628942489624, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.02469109185039997, + "learning_rate": 0.00033969494174772465, + "loss": 3.406843662261963, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.02417670376598835, + "learning_rate": 0.0003394303829840469, + "loss": 3.445046901702881, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.0237257182598114, + "learning_rate": 0.00033916579486906764, + "loss": 3.4490597248077393, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.02295520156621933, + "learning_rate": 0.0003389011776159607, + "loss": 3.4009876251220703, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.023158475756645203, + "learning_rate": 0.0003386365314379233, + "loss": 3.4383902549743652, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.023618478327989578, + "learning_rate": 0.0003383718565481763, + "loss": 3.421168804168701, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.022055329754948616, + "learning_rate": 0.0003381071531599633, + "loss": 3.382946729660034, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.021441346034407616, + "learning_rate": 0.00033784242148655115, + "loss": 3.4464712142944336, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.023031961172819138, + "learning_rate": 0.00033757766174122934, + "loss": 3.42844295501709, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.022832952439785004, + "learning_rate": 0.00033731287413731005, + "loss": 3.398231029510498, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.022094765678048134, + "learning_rate": 0.0003370480588881278, + "loss": 3.4270074367523193, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.0218376275151968, + "learning_rate": 0.0003367832162070395, + "loss": 3.4681742191314697, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.02364739589393139, + "learning_rate": 0.000336518346307424, + "loss": 3.431920051574707, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.023402225226163864, + "learning_rate": 0.0003362534494026824, + "loss": 3.4189023971557617, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.022775836288928986, + "learning_rate": 0.0003359885257062372, + "loss": 3.409396171569824, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.02445383556187153, + "learning_rate": 0.0003357235754315328, + "loss": 3.41017746925354, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.066352367401123, + "eval_runtime": 58.1658, + "eval_samples_per_second": 41.983, + "eval_steps_per_second": 1.324, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.02710847556591034, + "learning_rate": 0.0003354585987920345, + "loss": 3.3155131340026855, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.03428799286484718, + "learning_rate": 0.0003351935960012296, + "loss": 3.327408790588379, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.03434379771351814, + "learning_rate": 0.0003349285672726259, + "loss": 3.3282251358032227, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.03405233472585678, + "learning_rate": 0.0003346635128197522, + "loss": 3.2932448387145996, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.031194346025586128, + "learning_rate": 0.0003343984328561581, + "loss": 3.336615800857544, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.031816478818655014, + "learning_rate": 0.00033413332759541376, + "loss": 3.3375275135040283, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.03413299471139908, + "learning_rate": 0.0003338681972511098, + "loss": 3.3710694313049316, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.034200165420770645, + "learning_rate": 0.0003336030420368568, + "loss": 3.3248214721679688, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.030694160610437393, + "learning_rate": 0.0003333378621662857, + "loss": 3.3297901153564453, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.031172804534435272, + "learning_rate": 0.00033307265785304684, + "loss": 3.287217140197754, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.031914614140987396, + "learning_rate": 0.00033280742931081076, + "loss": 3.3218531608581543, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.0332055389881134, + "learning_rate": 0.00033254217675326737, + "loss": 3.3454785346984863, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.034956980496644974, + "learning_rate": 0.0003322769003941257, + "loss": 3.3193702697753906, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.03618055582046509, + "learning_rate": 0.00033201160044711423, + "loss": 3.3661017417907715, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.030322827398777008, + "learning_rate": 0.0003317462771259802, + "loss": 3.353989601135254, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.029768960550427437, + "learning_rate": 0.00033148093064449006, + "loss": 3.3254497051239014, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.028036782518029213, + "learning_rate": 0.0003312155612164284, + "loss": 3.3276920318603516, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.028774483129382133, + "learning_rate": 0.00033095016905559883, + "loss": 3.3458826541900635, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.030403515323996544, + "learning_rate": 0.0003306847543758227, + "loss": 3.3332793712615967, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.028468478471040726, + "learning_rate": 0.0003304193173909401, + "loss": 3.3167755603790283, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.029126476496458054, + "learning_rate": 0.00033015385831480873, + "loss": 3.299426794052124, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.030746638774871826, + "learning_rate": 0.0003298883773613043, + "loss": 3.3581809997558594, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.030241351574659348, + "learning_rate": 0.0003296228747443197, + "loss": 3.3415942192077637, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.029754415154457092, + "learning_rate": 0.0003293573506777659, + "loss": 3.3525662422180176, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.026896746829152107, + "learning_rate": 0.00032909180537557076, + "loss": 3.321808099746704, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.026361066848039627, + "learning_rate": 0.00032882623905167917, + "loss": 3.329258441925049, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.027747981250286102, + "learning_rate": 0.00032856065192005335, + "loss": 3.3774523735046387, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.026270100846886635, + "learning_rate": 0.00032829504419467194, + "loss": 3.3419976234436035, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.02731695957481861, + "learning_rate": 0.0003280294160895303, + "loss": 3.3068814277648926, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.026583299040794373, + "learning_rate": 0.0003277637678186402, + "loss": 3.303804874420166, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.026615949347615242, + "learning_rate": 0.00032749809959602973, + "loss": 3.3812599182128906, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.027221621945500374, + "learning_rate": 0.0003272324116357428, + "loss": 3.327211618423462, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.025025667622685432, + "learning_rate": 0.0003269667041518395, + "loss": 3.339163303375244, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.02593836560845375, + "learning_rate": 0.0003267009773583956, + "loss": 3.3517372608184814, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.023697076365351677, + "learning_rate": 0.00032643523146950234, + "loss": 3.379819869995117, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.02425902709364891, + "learning_rate": 0.0003261694666992664, + "loss": 3.390653610229492, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.025133250281214714, + "learning_rate": 0.0003259036832618096, + "loss": 3.364495277404785, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.025542905554175377, + "learning_rate": 0.0003256378813712688, + "loss": 3.3584814071655273, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.026041125878691673, + "learning_rate": 0.000325372061241796, + "loss": 3.36777400970459, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.02391037344932556, + "learning_rate": 0.00032510622308755746, + "loss": 3.3809385299682617, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.02379394695162773, + "learning_rate": 0.00032484036712273424, + "loss": 3.354257583618164, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.0234722550958395, + "learning_rate": 0.0003245744935615219, + "loss": 3.3736205101013184, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.024123650044202805, + "learning_rate": 0.0003243086026181296, + "loss": 3.3589959144592285, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.024013696238398552, + "learning_rate": 0.00032404269450678116, + "loss": 3.353874444961548, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.022617843002080917, + "learning_rate": 0.00032377676944171375, + "loss": 3.364056348800659, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.022730547934770584, + "learning_rate": 0.00032351082763717857, + "loss": 3.378067970275879, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.022257911041378975, + "learning_rate": 0.0003232448693074399, + "loss": 3.379342555999756, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.023247094824910164, + "learning_rate": 0.00032297889466677575, + "loss": 3.3537116050720215, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.022764021530747414, + "learning_rate": 0.000322712903929477, + "loss": 3.3969063758850098, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.02260403335094452, + "learning_rate": 0.0003224468973098477, + "loss": 3.3727574348449707, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.021772351115942, + "learning_rate": 0.0003221808750222044, + "loss": 3.335434675216675, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.02238459885120392, + "learning_rate": 0.0003219148372808766, + "loss": 3.3440277576446533, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.0230540968477726, + "learning_rate": 0.00032164878430020606, + "loss": 3.3997583389282227, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.0220821350812912, + "learning_rate": 0.00032138271629454684, + "loss": 3.377815008163452, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.021497417241334915, + "learning_rate": 0.00032111663347826505, + "loss": 3.3696370124816895, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.02163885347545147, + "learning_rate": 0.00032085053606573896, + "loss": 3.3489794731140137, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.022394269704818726, + "learning_rate": 0.0003205844242713584, + "loss": 3.3792812824249268, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.022390708327293396, + "learning_rate": 0.0003203182983095248, + "loss": 3.3733272552490234, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.023712158203125, + "learning_rate": 0.0003200521583946511, + "loss": 3.3716251850128174, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.02398860827088356, + "learning_rate": 0.0003197860047411613, + "loss": 3.381779670715332, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.02309161052107811, + "learning_rate": 0.00031951983756349066, + "loss": 3.3923277854919434, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.021400388330221176, + "learning_rate": 0.00031925365707608536, + "loss": 3.407397747039795, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.022534485906362534, + "learning_rate": 0.00031898746349340204, + "loss": 3.3740391731262207, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.023533202707767487, + "learning_rate": 0.0003187212570299082, + "loss": 3.3877944946289062, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.023398835211992264, + "learning_rate": 0.0003184550379000815, + "loss": 3.3936471939086914, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.020927201956510544, + "learning_rate": 0.00031818880631841005, + "loss": 3.339545249938965, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.021453995257616043, + "learning_rate": 0.0003179225624993915, + "loss": 3.3647522926330566, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.022847114130854607, + "learning_rate": 0.0003176563066575341, + "loss": 3.405144691467285, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.022766996175050735, + "learning_rate": 0.0003173900390073549, + "loss": 3.392810344696045, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.022103948518633842, + "learning_rate": 0.0003171237597633813, + "loss": 3.3973004817962646, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.022282181307673454, + "learning_rate": 0.0003168574691401496, + "loss": 3.409998893737793, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.025435522198677063, + "learning_rate": 0.0003165911673522053, + "loss": 3.423318386077881, + "step": 1800 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 7.67403340333056e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-1800/training_args.bin b/runs/l2r90-baseline/checkpoint-1800/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-1800/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-2160/chat_template.jinja b/runs/l2r90-baseline/checkpoint-2160/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-2160/config.json b/runs/l2r90-baseline/checkpoint-2160/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-2160/generation_config.json b/runs/l2r90-baseline/checkpoint-2160/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-2160/model.safetensors b/runs/l2r90-baseline/checkpoint-2160/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1d35be6fff00ef5890195bd2a404c4a944a2384d --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54a768ac68dc73a4c2e6245fde5293c2acab5fa980caaf4327bea8514ec149bf +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-2160/optimizer.pt b/runs/l2r90-baseline/checkpoint-2160/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..b9e2dbfc7f0043d744b78747b20804f613b986d5 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c0d3fae8c96a300056716e6c1013391ead11e35338cbf8189dce430f33361c4 +size 1166825338 diff --git a/runs/l2r90-baseline/checkpoint-2160/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-2160/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..88d2248182b7fded85e1613b19e4b23df99807a8 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08099d951fe225043bf5ed73f707acf33614689f66757b2a05711985ae68a49b +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2160/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-2160/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..f963be7f5cfeb7d06680588884ac6cd1c88f6843 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7de227da758404393630309e80adc97b4c158340b31e1f8ecfafc00d86ec519e +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2160/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-2160/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..45ba75ff4448dedf7ab83e4464cb19a4bce5fc27 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afa836b4fa8a014b4e440884fd4c2cf366a9b2474a622d72ac44bf2bedb57156 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2160/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-2160/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..181c4a277a2b37c89f2d06083f8b4a7ca6902792 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d94af1ab380b7f215a5ee3d3b50ab122adbd3dd0abfe48c29e83850925e0fa9 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2160/scheduler.pt b/runs/l2r90-baseline/checkpoint-2160/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..35aefe7762df83336d8bfddab210c09c85e6b3aa --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7d6a0a5c94f665a32f55d7b034b4fc829f9f44c468d851ff5b24b22d8388e63 +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-2160/tokenizer.json b/runs/l2r90-baseline/checkpoint-2160/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-2160/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-2160/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-2160/trainer_state.json b/runs/l2r90-baseline/checkpoint-2160/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..5ab7f7d0ede6076cca695eb7007edc123b7fce0e --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/trainer_state.json @@ -0,0 +1,15386 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 30.0, + "eval_steps": 500, + "global_step": 2160, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 5.549652099609375, + "eval_runtime": 57.761, + "eval_samples_per_second": 42.278, + "eval_steps_per_second": 1.333, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.03958690166473389, + "learning_rate": 0.0005919487226162231, + "loss": 5.459286689758301, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.03299890458583832, + "learning_rate": 0.0005918869546699877, + "loss": 5.45464563369751, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.028040863573551178, + "learning_rate": 0.0005918249539727509, + "loss": 5.444796562194824, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.028365720063447952, + "learning_rate": 0.0005917627205744655, + "loss": 5.447852611541748, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.02822224237024784, + "learning_rate": 0.0005917002545252719, + "loss": 5.453883171081543, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.036075398325920105, + "learning_rate": 0.0005916375558754977, + "loss": 5.462671279907227, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04683876410126686, + "learning_rate": 0.0005915746246756581, + "loss": 5.443917274475098, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.05457588657736778, + "learning_rate": 0.0005915114609764558, + "loss": 5.444374084472656, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.04250836744904518, + "learning_rate": 0.0005914480648287804, + "loss": 5.435029983520508, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.029477844014763832, + "learning_rate": 0.0005913844362837093, + "loss": 5.421862602233887, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.03673284500837326, + "learning_rate": 0.0005913205753925066, + "loss": 5.430391311645508, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.030489861965179443, + "learning_rate": 0.0005912564822066241, + "loss": 5.389314651489258, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.04037046805024147, + "learning_rate": 0.0005911921567777005, + "loss": 5.425060272216797, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.0445217490196228, + "learning_rate": 0.0005911275991575613, + "loss": 5.437565326690674, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.03885014355182648, + "learning_rate": 0.0005910628093982198, + "loss": 5.410930633544922, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.0334964245557785, + "learning_rate": 0.0005909977875518759, + "loss": 5.412092208862305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.03499286621809006, + "learning_rate": 0.0005909325336709164, + "loss": 5.382216453552246, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.04171644523739815, + "learning_rate": 0.0005908670478079152, + "loss": 5.387385368347168, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.057558078318834305, + "learning_rate": 0.000590801330015633, + "loss": 5.437925815582275, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.053846780210733414, + "learning_rate": 0.0005907353803470177, + "loss": 5.41644811630249, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.027259036898612976, + "learning_rate": 0.0005906691988552034, + "loss": 5.409185409545898, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.03394043818116188, + "learning_rate": 0.0005906027855935115, + "loss": 5.377973556518555, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.03731574863195419, + "learning_rate": 0.0005905361406154501, + "loss": 5.382505893707275, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.04705141484737396, + "learning_rate": 0.0005904692639747137, + "loss": 5.413661003112793, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.044811323285102844, + "learning_rate": 0.0005904021557251837, + "loss": 5.417835235595703, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.04000752419233322, + "learning_rate": 0.0005903348159209277, + "loss": 5.406160354614258, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.038364164531230927, + "learning_rate": 0.0005902672446162007, + "loss": 5.357950687408447, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.030500127002596855, + "learning_rate": 0.0005901994418654432, + "loss": 5.344951629638672, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.030644729733467102, + "learning_rate": 0.0005901314077232829, + "loss": 5.3599348068237305, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.037976790219545364, + "learning_rate": 0.0005900631422445335, + "loss": 5.354084491729736, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.04106011986732483, + "learning_rate": 0.0005899946454841955, + "loss": 5.363678455352783, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.04334832355380058, + "learning_rate": 0.0005899259174974553, + "loss": 5.350959300994873, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.049939028918743134, + "learning_rate": 0.000589856958339686, + "loss": 5.3317155838012695, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.05300058051943779, + "learning_rate": 0.0005897877680664465, + "loss": 5.353314399719238, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.05943184345960617, + "learning_rate": 0.0005897183467334822, + "loss": 5.3683929443359375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.05600789934396744, + "learning_rate": 0.0005896486943967249, + "loss": 5.345516204833984, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.05392718315124512, + "learning_rate": 0.0005895788111122919, + "loss": 5.291389465332031, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.037913475185632706, + "learning_rate": 0.0005895086969364871, + "loss": 5.353423118591309, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.03362123295664787, + "learning_rate": 0.0005894383519258001, + "loss": 5.379500389099121, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032558735460042953, + "learning_rate": 0.0005893677761369066, + "loss": 5.34043025970459, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.04074402526021004, + "learning_rate": 0.0005892969696266683, + "loss": 5.303869247436523, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.04270484298467636, + "learning_rate": 0.0005892259324521328, + "loss": 5.27297830581665, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.04604959115386009, + "learning_rate": 0.0005891546646705334, + "loss": 5.313268184661865, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.04462246224284172, + "learning_rate": 0.0005890831663392891, + "loss": 5.31167459487915, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.04966094344854355, + "learning_rate": 0.0005890114375160052, + "loss": 5.316695690155029, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.04662073776125908, + "learning_rate": 0.0005889394782584718, + "loss": 5.316898822784424, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.04380315542221069, + "learning_rate": 0.0005888672886246656, + "loss": 5.285087585449219, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03239579498767853, + "learning_rate": 0.0005887948686727483, + "loss": 5.304283142089844, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.029690183699131012, + "learning_rate": 0.0005887222184610675, + "loss": 5.305715084075928, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.03444967418909073, + "learning_rate": 0.0005886493380481559, + "loss": 5.2707672119140625, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.03302565962076187, + "learning_rate": 0.0005885762274927322, + "loss": 5.277251243591309, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.028477763757109642, + "learning_rate": 0.0005885028868537, + "loss": 5.295318603515625, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.022988801822066307, + "learning_rate": 0.0005884293161901487, + "loss": 5.249057292938232, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.026115216314792633, + "learning_rate": 0.000588355515561353, + "loss": 5.270636081695557, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.02576332725584507, + "learning_rate": 0.0005882814850267723, + "loss": 5.263096332550049, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.023978352546691895, + "learning_rate": 0.0005882072246460521, + "loss": 5.276605606079102, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.020695513114333153, + "learning_rate": 0.0005881327344790223, + "loss": 5.2141432762146, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01952219195663929, + "learning_rate": 0.0005880580145856986, + "loss": 5.214289665222168, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.022633206099271774, + "learning_rate": 0.0005879830650262813, + "loss": 5.236690998077393, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.023082347586750984, + "learning_rate": 0.0005879078858611557, + "loss": 5.2092156410217285, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.019620463252067566, + "learning_rate": 0.0005878324771508927, + "loss": 5.264451503753662, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.021504493430256844, + "learning_rate": 0.0005877568389562475, + "loss": 5.23444938659668, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.025672364979982376, + "learning_rate": 0.0005876809713381606, + "loss": 5.205289363861084, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.034436631947755814, + "learning_rate": 0.0005876048743577569, + "loss": 5.2047648429870605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.042389530688524246, + "learning_rate": 0.0005875285480763466, + "loss": 5.192008018493652, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.03310716897249222, + "learning_rate": 0.0005874519925554244, + "loss": 5.162734031677246, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.03029288910329342, + "learning_rate": 0.0005873752078566694, + "loss": 5.212151527404785, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04619310051202774, + "learning_rate": 0.0005872981940419459, + "loss": 5.215909004211426, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.05618087574839592, + "learning_rate": 0.0005872209511733027, + "loss": 5.210192680358887, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.03783707693219185, + "learning_rate": 0.0005871434793129726, + "loss": 5.201597690582275, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.03721104562282562, + "learning_rate": 0.0005870657785233735, + "loss": 5.188891410827637, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.033336155116558075, + "learning_rate": 0.0005869878488671075, + "loss": 5.1598615646362305, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.2592363357543945, + "eval_runtime": 57.3343, + "eval_samples_per_second": 42.592, + "eval_steps_per_second": 1.343, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03719741478562355, + "learning_rate": 0.0005869096904069611, + "loss": 5.1646623611450195, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.04888090118765831, + "learning_rate": 0.0005868313032059052, + "loss": 5.171182155609131, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.05073149502277374, + "learning_rate": 0.0005867526873270949, + "loss": 5.183949947357178, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.04187731072306633, + "learning_rate": 0.0005866738428338695, + "loss": 5.176133155822754, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.039445552974939346, + "learning_rate": 0.000586594769789753, + "loss": 5.148100852966309, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.035551708191633224, + "learning_rate": 0.0005865154682584524, + "loss": 5.151512145996094, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.031175289303064346, + "learning_rate": 0.0005864359383038602, + "loss": 5.102468967437744, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.03696789592504501, + "learning_rate": 0.000586356179990052, + "loss": 5.119902610778809, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.04836135357618332, + "learning_rate": 0.0005862761933812875, + "loss": 5.15183162689209, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.05206914246082306, + "learning_rate": 0.0005861959785420106, + "loss": 5.154870986938477, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.046368084847927094, + "learning_rate": 0.000586115535536849, + "loss": 5.130651473999023, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.07146432995796204, + "learning_rate": 0.000586034864430614, + "loss": 5.175804138183594, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.08809789270162582, + "learning_rate": 0.0005859539652883008, + "loss": 5.229716777801514, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.08116360008716583, + "learning_rate": 0.0005858728381750888, + "loss": 5.21910285949707, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.08282570540904999, + "learning_rate": 0.0005857914831563402, + "loss": 5.2336506843566895, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.07403547316789627, + "learning_rate": 0.0005857099002976016, + "loss": 5.251604080200195, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.07024004310369492, + "learning_rate": 0.0005856280896646026, + "loss": 5.267114162445068, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.056330062448978424, + "learning_rate": 0.0005855460513232567, + "loss": 5.201976299285889, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.051763955503702164, + "learning_rate": 0.0005854637853396606, + "loss": 5.1733198165893555, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.039900992065668106, + "learning_rate": 0.0005853812917800945, + "loss": 5.222941875457764, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.04187863692641258, + "learning_rate": 0.0005852985707110221, + "loss": 5.165033340454102, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.042645812034606934, + "learning_rate": 0.0005852156221990901, + "loss": 5.157554626464844, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.03856075555086136, + "learning_rate": 0.0005851324463111289, + "loss": 5.184629440307617, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.029387330636382103, + "learning_rate": 0.0005850490431141516, + "loss": 5.118739128112793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.027584781870245934, + "learning_rate": 0.0005849654126753545, + "loss": 5.100855350494385, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.023655924946069717, + "learning_rate": 0.0005848815550621175, + "loss": 5.154520034790039, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.024163050577044487, + "learning_rate": 0.0005847974703420028, + "loss": 5.105189323425293, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.020369920879602432, + "learning_rate": 0.0005847131585827564, + "loss": 5.113671779632568, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.021030737087130547, + "learning_rate": 0.0005846286198523061, + "loss": 5.0847649574279785, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.01729249767959118, + "learning_rate": 0.0005845438542187638, + "loss": 5.053225517272949, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.018704712390899658, + "learning_rate": 0.0005844588617504236, + "loss": 5.079106330871582, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.02049064077436924, + "learning_rate": 0.0005843736425157621, + "loss": 5.044185161590576, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.020736856386065483, + "learning_rate": 0.000584288196583439, + "loss": 5.041932106018066, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.016543276607990265, + "learning_rate": 0.0005842025240222966, + "loss": 5.032522201538086, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.018414106220006943, + "learning_rate": 0.0005841166249013598, + "loss": 5.037397384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.01611698977649212, + "learning_rate": 0.0005840304992898359, + "loss": 5.038179397583008, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.014310561120510101, + "learning_rate": 0.0005839441472571147, + "loss": 5.0146484375, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.015472214668989182, + "learning_rate": 0.0005838575688727685, + "loss": 5.060323715209961, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.016373401507735252, + "learning_rate": 0.0005837707642065518, + "loss": 5.032100677490234, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.018694665282964706, + "learning_rate": 0.000583683733328402, + "loss": 5.036993026733398, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.019977087154984474, + "learning_rate": 0.0005835964763084378, + "loss": 4.991037368774414, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.023741086944937706, + "learning_rate": 0.000583508993216961, + "loss": 5.043832778930664, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.025061562657356262, + "learning_rate": 0.0005834212841244548, + "loss": 5.016516208648682, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.020848384127020836, + "learning_rate": 0.0005833333491015851, + "loss": 4.986289978027344, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.015086354687809944, + "learning_rate": 0.0005832451882191995, + "loss": 4.9948625564575195, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.015096964314579964, + "learning_rate": 0.0005831568015483274, + "loss": 4.996342658996582, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016932200640439987, + "learning_rate": 0.0005830681891601807, + "loss": 4.998758316040039, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.019936930388212204, + "learning_rate": 0.0005829793511261525, + "loss": 4.959854602813721, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.027658682316541672, + "learning_rate": 0.0005828902875178183, + "loss": 4.968629837036133, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.03299275413155556, + "learning_rate": 0.0005828009984069347, + "loss": 4.994814872741699, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.033717650920152664, + "learning_rate": 0.0005827114838654406, + "loss": 4.976295471191406, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.02969924360513687, + "learning_rate": 0.0005826217439654562, + "loss": 4.969155788421631, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.02846827358007431, + "learning_rate": 0.0005825317787792831, + "loss": 4.989721298217773, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.03670600801706314, + "learning_rate": 0.0005824415883794049, + "loss": 4.988689422607422, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.045570675283670425, + "learning_rate": 0.0005823511728384863, + "loss": 4.974407196044922, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.04564111679792404, + "learning_rate": 0.0005822605322293733, + "loss": 4.966331481933594, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.04173210263252258, + "learning_rate": 0.0005821696666250937, + "loss": 4.925197601318359, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.04228242486715317, + "learning_rate": 0.0005820785760988559, + "loss": 4.965020656585693, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.05144782364368439, + "learning_rate": 0.0005819872607240503, + "loss": 5.0022149085998535, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.049400657415390015, + "learning_rate": 0.0005818957205742478, + "loss": 4.9635114669799805, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.043692708015441895, + "learning_rate": 0.0005818039557232005, + "loss": 4.977723121643066, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.04891742020845413, + "learning_rate": 0.0005817119662448421, + "loss": 4.996312141418457, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.035564981400966644, + "learning_rate": 0.0005816197522132866, + "loss": 4.940932273864746, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04679827764630318, + "learning_rate": 0.0005815273137028292, + "loss": 4.983131408691406, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.05489890277385712, + "learning_rate": 0.0005814346507879459, + "loss": 4.977827548980713, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04640809819102287, + "learning_rate": 0.0005813417635432937, + "loss": 4.954761505126953, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04439452290534973, + "learning_rate": 0.00058124865204371, + "loss": 4.914865970611572, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.04587096720933914, + "learning_rate": 0.0005811553163642131, + "loss": 4.949592590332031, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.044119734317064285, + "learning_rate": 0.0005810617565800019, + "loss": 4.956068992614746, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.043133001774549484, + "learning_rate": 0.0005809679727664559, + "loss": 4.993522644042969, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04315278306603432, + "learning_rate": 0.0005808739649991348, + "loss": 4.894291877746582, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04819472134113312, + "learning_rate": 0.0005807797333537792, + "loss": 4.958306789398193, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.023706436157227, + "eval_runtime": 57.4082, + "eval_samples_per_second": 42.537, + "eval_steps_per_second": 1.341, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.04403403028845787, + "learning_rate": 0.0005806852779063098, + "loss": 4.884971618652344, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.041993748396635056, + "learning_rate": 0.0005805905987328275, + "loss": 4.916353225708008, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.042302004992961884, + "learning_rate": 0.0005804956959096138, + "loss": 4.9012556076049805, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.044500596821308136, + "learning_rate": 0.0005804005695131299, + "loss": 4.915240287780762, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.03914736211299896, + "learning_rate": 0.0005803052196200177, + "loss": 4.885915756225586, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.0361703559756279, + "learning_rate": 0.0005802096463070988, + "loss": 4.829287528991699, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03432302549481392, + "learning_rate": 0.0005801138496513749, + "loss": 4.926834583282471, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.0317804180085659, + "learning_rate": 0.0005800178297300276, + "loss": 4.8707170486450195, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.035261351615190506, + "learning_rate": 0.0005799215866204183, + "loss": 4.897791862487793, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.03276067599654198, + "learning_rate": 0.0005798251204000886, + "loss": 4.864504814147949, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.034757066518068314, + "learning_rate": 0.0005797284311467594, + "loss": 4.890812873840332, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.033726081252098083, + "learning_rate": 0.0005796315189383316, + "loss": 4.831684112548828, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.03074726276099682, + "learning_rate": 0.0005795343838528855, + "loss": 4.8534111976623535, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02215893566608429, + "learning_rate": 0.0005794370259686811, + "loss": 4.8311238288879395, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.02132033370435238, + "learning_rate": 0.0005793394453641581, + "loss": 4.842617034912109, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020825816318392754, + "learning_rate": 0.0005792416421179351, + "loss": 4.84535026550293, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.019334357231855392, + "learning_rate": 0.0005791436163088106, + "loss": 4.839937686920166, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017372334375977516, + "learning_rate": 0.0005790453680157622, + "loss": 4.8115553855896, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.0190406683832407, + "learning_rate": 0.0005789468973179468, + "loss": 4.832542419433594, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.017372574657201767, + "learning_rate": 0.0005788482042947004, + "loss": 4.793632507324219, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.019171632826328278, + "learning_rate": 0.0005787492890255382, + "loss": 4.776987552642822, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.02109278179705143, + "learning_rate": 0.0005786501515901545, + "loss": 4.8216657638549805, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.026234494522213936, + "learning_rate": 0.0005785507920684225, + "loss": 4.8109049797058105, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.0357641875743866, + "learning_rate": 0.0005784512105403944, + "loss": 4.821109771728516, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.03627947345376015, + "learning_rate": 0.000578351407086301, + "loss": 4.810632228851318, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.032253146171569824, + "learning_rate": 0.0005782513817865527, + "loss": 4.7930521965026855, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.03402210772037506, + "learning_rate": 0.0005781511347217375, + "loss": 4.772500991821289, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.0321584977209568, + "learning_rate": 0.000578050665972623, + "loss": 4.781399726867676, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.030242707580327988, + "learning_rate": 0.0005779499756201549, + "loss": 4.802542686462402, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.02361336164176464, + "learning_rate": 0.0005778490637454576, + "loss": 4.771912574768066, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.028781302273273468, + "learning_rate": 0.000577747930429834, + "loss": 4.796924591064453, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.027042683213949203, + "learning_rate": 0.0005776465757547654, + "loss": 4.745218276977539, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.02715984545648098, + "learning_rate": 0.0005775449998019113, + "loss": 4.792944431304932, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.0314096100628376, + "learning_rate": 0.00057744320265311, + "loss": 4.7649993896484375, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.028749193996191025, + "learning_rate": 0.0005773411843903773, + "loss": 4.7980546951293945, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.02655309997498989, + "learning_rate": 0.0005772389450959075, + "loss": 4.746520519256592, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.027497397735714912, + "learning_rate": 0.000577136484852073, + "loss": 4.7490715980529785, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.0343393050134182, + "learning_rate": 0.000577033803741424, + "loss": 4.74436092376709, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.04444412514567375, + "learning_rate": 0.0005769309018466891, + "loss": 4.755973815917969, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.0437556728720665, + "learning_rate": 0.0005768277792507744, + "loss": 4.784368515014648, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0369921550154686, + "learning_rate": 0.0005767244360367638, + "loss": 4.753410339355469, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.05038141459226608, + "learning_rate": 0.0005766208722879192, + "loss": 4.755491256713867, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.04511422663927078, + "learning_rate": 0.00057651708808768, + "loss": 4.760293006896973, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.048762883991003036, + "learning_rate": 0.0005764130835196632, + "loss": 4.755958080291748, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.06749686598777771, + "learning_rate": 0.0005763088586676634, + "loss": 4.802557468414307, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.06451250612735748, + "learning_rate": 0.0005762044136156528, + "loss": 4.812925338745117, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.05436156690120697, + "learning_rate": 0.0005760997484477809, + "loss": 4.807103157043457, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.05233754590153694, + "learning_rate": 0.0005759948632483742, + "loss": 4.782449722290039, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.05519746616482735, + "learning_rate": 0.0005758897581019371, + "loss": 4.808449745178223, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.0578567236661911, + "learning_rate": 0.000575784433093151, + "loss": 4.7676005363464355, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.0588264986872673, + "learning_rate": 0.0005756788883068743, + "loss": 4.805377960205078, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04225790873169899, + "learning_rate": 0.0005755731238281423, + "loss": 4.748775005340576, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.0376368910074234, + "learning_rate": 0.0005754671397421678, + "loss": 4.759791374206543, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03830771893262863, + "learning_rate": 0.0005753609361343402, + "loss": 4.729974746704102, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.03684520348906517, + "learning_rate": 0.0005752545130902256, + "loss": 4.729854583740234, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02983408235013485, + "learning_rate": 0.0005751478706955674, + "loss": 4.78502082824707, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02854904904961586, + "learning_rate": 0.0005750410090362854, + "loss": 4.717472076416016, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.023819150403141975, + "learning_rate": 0.0005749339281984761, + "loss": 4.717124938964844, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.02477158047258854, + "learning_rate": 0.0005748266282684124, + "loss": 4.71781063079834, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.022002534940838814, + "learning_rate": 0.0005747191093325443, + "loss": 4.7480669021606445, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.02133483812212944, + "learning_rate": 0.0005746113714774976, + "loss": 4.693770408630371, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.022407138720154762, + "learning_rate": 0.0005745034147900747, + "loss": 4.720734119415283, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.020850926637649536, + "learning_rate": 0.0005743952393572544, + "loss": 4.745505332946777, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.020420582965016365, + "learning_rate": 0.0005742868452661918, + "loss": 4.688610076904297, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.019014107063412666, + "learning_rate": 0.0005741782326042181, + "loss": 4.673436164855957, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.017393039539456367, + "learning_rate": 0.0005740694014588403, + "loss": 4.6941118240356445, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.017877329140901566, + "learning_rate": 0.0005739603519177419, + "loss": 4.667474269866943, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.017749635502696037, + "learning_rate": 0.0005738510840687821, + "loss": 4.6901092529296875, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.018746376037597656, + "learning_rate": 0.000573741597999996, + "loss": 4.680166721343994, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016260217875242233, + "learning_rate": 0.0005736318937995947, + "loss": 4.6799821853637695, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015099814161658287, + "learning_rate": 0.0005735219715559646, + "loss": 4.680666923522949, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.016539735719561577, + "learning_rate": 0.0005734118313576683, + "loss": 4.639509201049805, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 4.768754482269287, + "eval_runtime": 58.3352, + "eval_samples_per_second": 41.862, + "eval_steps_per_second": 1.32, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.019733518362045288, + "learning_rate": 0.0005733014732934436, + "loss": 4.635312080383301, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.022571036592125893, + "learning_rate": 0.0005731908974522042, + "loss": 4.6041765213012695, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.022238552570343018, + "learning_rate": 0.0005730801039230389, + "loss": 4.648976802825928, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.02115180902183056, + "learning_rate": 0.0005729690927952123, + "loss": 4.623968124389648, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.021789928898215294, + "learning_rate": 0.0005728578641581637, + "loss": 4.620028018951416, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.02593887411057949, + "learning_rate": 0.0005727464181015081, + "loss": 4.591062545776367, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.0347842313349247, + "learning_rate": 0.0005726347547150357, + "loss": 4.632149696350098, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04119519516825676, + "learning_rate": 0.0005725228740887117, + "loss": 4.6254191398620605, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.04895750433206558, + "learning_rate": 0.0005724107763126761, + "loss": 4.621372222900391, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.05752161890268326, + "learning_rate": 0.0005722984614772442, + "loss": 4.61965274810791, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05115870013833046, + "learning_rate": 0.000572185929672906, + "loss": 4.63119649887085, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.05721089243888855, + "learning_rate": 0.0005720731809903263, + "loss": 4.652647018432617, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.0646839588880539, + "learning_rate": 0.000571960215520345, + "loss": 4.642184257507324, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.0682854950428009, + "learning_rate": 0.0005718470333539757, + "loss": 4.676589012145996, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.057402849197387695, + "learning_rate": 0.0005717336345824077, + "loss": 4.6585283279418945, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.05885373800992966, + "learning_rate": 0.0005716200192970043, + "loss": 4.671728134155273, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.05265705659985542, + "learning_rate": 0.0005715061875893032, + "loss": 4.641249656677246, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.05417842045426369, + "learning_rate": 0.0005713921395510166, + "loss": 4.661381721496582, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.06602369248867035, + "learning_rate": 0.0005712778752740307, + "loss": 4.684137344360352, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.06004182621836662, + "learning_rate": 0.0005711633948504066, + "loss": 4.625332832336426, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.04801511764526367, + "learning_rate": 0.0005710486983723787, + "loss": 4.664791107177734, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.04380105435848236, + "learning_rate": 0.0005709337859323561, + "loss": 4.664966583251953, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.03858727216720581, + "learning_rate": 0.0005708186576229218, + "loss": 4.661867618560791, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03240576758980751, + "learning_rate": 0.0005707033135368323, + "loss": 4.5855326652526855, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.028561925515532494, + "learning_rate": 0.0005705877537670184, + "loss": 4.637331008911133, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.026389246806502342, + "learning_rate": 0.0005704719784065846, + "loss": 4.598666667938232, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.02484819106757641, + "learning_rate": 0.0005703559875488088, + "loss": 4.601729393005371, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.02522977627813816, + "learning_rate": 0.0005702397812871429, + "loss": 4.609833717346191, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02589508146047592, + "learning_rate": 0.0005701233597152121, + "loss": 4.574592590332031, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.023374728858470917, + "learning_rate": 0.0005700067229268154, + "loss": 4.61362361907959, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.02465851791203022, + "learning_rate": 0.0005698898710159245, + "loss": 4.639899253845215, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.02258715033531189, + "learning_rate": 0.0005697728040766852, + "loss": 4.5656538009643555, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.018009789288043976, + "learning_rate": 0.0005696555222034162, + "loss": 4.61449670791626, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02121281623840332, + "learning_rate": 0.0005695380254906094, + "loss": 4.553830623626709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.018078701570630074, + "learning_rate": 0.0005694203140329296, + "loss": 4.534579277038574, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.0169303547590971, + "learning_rate": 0.000569302387925215, + "loss": 4.533995628356934, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.016318142414093018, + "learning_rate": 0.0005691842472624764, + "loss": 4.544438362121582, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.016664575785398483, + "learning_rate": 0.0005690658921398977, + "loss": 4.5913286209106445, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.01641504280269146, + "learning_rate": 0.0005689473226528354, + "loss": 4.520743370056152, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.017012735828757286, + "learning_rate": 0.0005688285388968187, + "loss": 4.517453193664551, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01672312617301941, + "learning_rate": 0.0005687095409675499, + "loss": 4.5235915184021, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.016893865540623665, + "learning_rate": 0.000568590328960903, + "loss": 4.543200492858887, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.017530135810375214, + "learning_rate": 0.0005684709029729253, + "loss": 4.54912805557251, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.018698599189519882, + "learning_rate": 0.0005683512630998361, + "loss": 4.522650718688965, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01984071172773838, + "learning_rate": 0.000568231409438027, + "loss": 4.546173095703125, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.023218289017677307, + "learning_rate": 0.0005681113420840619, + "loss": 4.5305094718933105, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.028968270868062973, + "learning_rate": 0.000567991061134677, + "loss": 4.521440505981445, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.032370273023843765, + "learning_rate": 0.0005678705666867805, + "loss": 4.54168701171875, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.033967748284339905, + "learning_rate": 0.0005677498588374524, + "loss": 4.538717269897461, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.028757929801940918, + "learning_rate": 0.0005676289376839452, + "loss": 4.517228603363037, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.029757825657725334, + "learning_rate": 0.0005675078033236827, + "loss": 4.510310173034668, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.03581003099679947, + "learning_rate": 0.0005673864558542605, + "loss": 4.506901264190674, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.03483985736966133, + "learning_rate": 0.0005672648953734462, + "loss": 4.549284934997559, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.031016027554869652, + "learning_rate": 0.000567143121979179, + "loss": 4.530084609985352, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.029373526573181152, + "learning_rate": 0.0005670211357695693, + "loss": 4.492218971252441, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.031661372631788254, + "learning_rate": 0.0005668989368428994, + "loss": 4.502824783325195, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.027251387014985085, + "learning_rate": 0.0005667765252976227, + "loss": 4.49193811416626, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02263176068663597, + "learning_rate": 0.0005666539012323639, + "loss": 4.512019157409668, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.02355228364467621, + "learning_rate": 0.0005665310647459189, + "loss": 4.508917331695557, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.023297371342778206, + "learning_rate": 0.0005664080159372551, + "loss": 4.497771263122559, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.019364114850759506, + "learning_rate": 0.0005662847549055107, + "loss": 4.51470947265625, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.023899901658296585, + "learning_rate": 0.0005661612817499947, + "loss": 4.48771858215332, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.02330983616411686, + "learning_rate": 0.0005660375965701872, + "loss": 4.478496551513672, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.027252234518527985, + "learning_rate": 0.0005659136994657392, + "loss": 4.497672080993652, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.030886363238096237, + "learning_rate": 0.0005657895905364723, + "loss": 4.49789571762085, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02515559457242489, + "learning_rate": 0.0005656652698823788, + "loss": 4.490771293640137, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020107125863432884, + "learning_rate": 0.0005655407376036217, + "loss": 4.481879234313965, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.022076716646552086, + "learning_rate": 0.0005654159938005345, + "loss": 4.494697570800781, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.025525815784931183, + "learning_rate": 0.0005652910385736207, + "loss": 4.517453193664551, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.03005894646048546, + "learning_rate": 0.0005651658720235548, + "loss": 4.453815460205078, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.030178502202033997, + "learning_rate": 0.0005650404942511812, + "loss": 4.493922233581543, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.029249394312500954, + "learning_rate": 0.0005649149053575141, + "loss": 4.453763008117676, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 4.607339382171631, + "eval_runtime": 57.9216, + "eval_samples_per_second": 42.16, + "eval_steps_per_second": 1.329, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.03362950682640076, + "learning_rate": 0.0005647891054437386, + "loss": 4.430700302124023, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.04223684221506119, + "learning_rate": 0.0005646630946112096, + "loss": 4.486915588378906, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.05192171037197113, + "learning_rate": 0.0005645368729614512, + "loss": 4.467541694641113, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.042525868862867355, + "learning_rate": 0.0005644104405961582, + "loss": 4.45252799987793, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.0406370609998703, + "learning_rate": 0.0005642837976171949, + "loss": 4.414548873901367, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.044476233422756195, + "learning_rate": 0.0005641569441265952, + "loss": 4.453492164611816, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.04687326401472092, + "learning_rate": 0.0005640298802265626, + "loss": 4.477725028991699, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.04011611267924309, + "learning_rate": 0.0005639026060194704, + "loss": 4.423356533050537, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.04971884936094284, + "learning_rate": 0.0005637751216078609, + "loss": 4.4853515625, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.05863234028220177, + "learning_rate": 0.0005636474270944461, + "loss": 4.466001033782959, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.057710036635398865, + "learning_rate": 0.0005635195225821072, + "loss": 4.466723442077637, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.05155074968934059, + "learning_rate": 0.0005633914081738945, + "loss": 4.4823198318481445, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.04535634070634842, + "learning_rate": 0.0005632630839730275, + "loss": 4.471219062805176, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.045147854834795, + "learning_rate": 0.0005631345500828946, + "loss": 4.476888179779053, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.03730619326233864, + "learning_rate": 0.0005630058066070533, + "loss": 4.428932189941406, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.034599728882312775, + "learning_rate": 0.0005628768536492299, + "loss": 4.47627592086792, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.032904356718063354, + "learning_rate": 0.0005627476913133193, + "loss": 4.422671794891357, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.03356337174773216, + "learning_rate": 0.0005626183197033856, + "loss": 4.456406593322754, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.035042595118284225, + "learning_rate": 0.0005624887389236609, + "loss": 4.371820449829102, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.03269032761454582, + "learning_rate": 0.0005623589490785462, + "loss": 4.402605056762695, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.028261031955480576, + "learning_rate": 0.0005622289502726108, + "loss": 4.433136940002441, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03301700949668884, + "learning_rate": 0.0005620987426105925, + "loss": 4.388359546661377, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.031664177775382996, + "learning_rate": 0.000561968326197397, + "loss": 4.41859245300293, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.026281699538230896, + "learning_rate": 0.0005618377011380988, + "loss": 4.406128883361816, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.023097911849617958, + "learning_rate": 0.0005617068675379398, + "loss": 4.389717102050781, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.022349286824464798, + "learning_rate": 0.0005615758255023306, + "loss": 4.408173561096191, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.021958226338028908, + "learning_rate": 0.0005614445751368491, + "loss": 4.434830188751221, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024551713839173317, + "learning_rate": 0.0005613131165472415, + "loss": 4.386944770812988, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.028048967942595482, + "learning_rate": 0.0005611814498394216, + "loss": 4.402279376983643, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02733907848596573, + "learning_rate": 0.0005610495751194708, + "loss": 4.403069496154785, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.020858559757471085, + "learning_rate": 0.0005609174924936384, + "loss": 4.406920909881592, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.021491149440407753, + "learning_rate": 0.0005607852020683409, + "loss": 4.385799407958984, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.021495329216122627, + "learning_rate": 0.0005606527039501621, + "loss": 4.4252495765686035, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.019708355888724327, + "learning_rate": 0.0005605199982458535, + "loss": 4.388864994049072, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.01886274851858616, + "learning_rate": 0.0005603870850623338, + "loss": 4.335962295532227, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.01955341547727585, + "learning_rate": 0.0005602539645066884, + "loss": 4.373560905456543, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.018962012603878975, + "learning_rate": 0.0005601206366861706, + "loss": 4.380022048950195, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.016663335263729095, + "learning_rate": 0.0005599871017081999, + "loss": 4.421034336090088, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01676873490214348, + "learning_rate": 0.0005598533596803631, + "loss": 4.382775783538818, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.014379126951098442, + "learning_rate": 0.0005597194107104137, + "loss": 4.41373872756958, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.015897583216428757, + "learning_rate": 0.000559585254906272, + "loss": 4.377284049987793, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.016513120383024216, + "learning_rate": 0.000559450892376025, + "loss": 4.41200065612793, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.016625918447971344, + "learning_rate": 0.000559316323227926, + "loss": 4.381187438964844, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.016028188169002533, + "learning_rate": 0.000559181547570395, + "loss": 4.353909015655518, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01682914048433304, + "learning_rate": 0.0005590465655120183, + "loss": 4.382326126098633, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.017078889533877373, + "learning_rate": 0.0005589113771615486, + "loss": 4.334949493408203, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.018182938918471336, + "learning_rate": 0.0005587759826279046, + "loss": 4.348130226135254, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01940886303782463, + "learning_rate": 0.0005586403820201713, + "loss": 4.339404106140137, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.020644692704081535, + "learning_rate": 0.0005585045754475995, + "loss": 4.358404159545898, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02433430217206478, + "learning_rate": 0.0005583685630196064, + "loss": 4.3660502433776855, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.027824537828564644, + "learning_rate": 0.0005582323448457744, + "loss": 4.34050178527832, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.02973582223057747, + "learning_rate": 0.0005580959210358523, + "loss": 4.348322868347168, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.030396128073334694, + "learning_rate": 0.0005579592916997542, + "loss": 4.356224060058594, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.029760530218482018, + "learning_rate": 0.0005578224569475599, + "loss": 4.357470989227295, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.02855013869702816, + "learning_rate": 0.0005576854168895147, + "loss": 4.335353374481201, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03376635164022446, + "learning_rate": 0.0005575481716360292, + "loss": 4.331794738769531, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.034431129693984985, + "learning_rate": 0.0005574107212976796, + "loss": 4.336880207061768, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.03524135425686836, + "learning_rate": 0.000557273065985207, + "loss": 4.31537389755249, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.03152058273553848, + "learning_rate": 0.0005571352058095179, + "loss": 4.37662410736084, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02878640592098236, + "learning_rate": 0.0005569971408816838, + "loss": 4.292388916015625, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.029663654044270515, + "learning_rate": 0.0005568588713129409, + "loss": 4.344346523284912, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.027895759791135788, + "learning_rate": 0.0005567203972146906, + "loss": 4.365072727203369, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.026465769857168198, + "learning_rate": 0.000556581718698499, + "loss": 4.334959506988525, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.03191852569580078, + "learning_rate": 0.000556442835876097, + "loss": 4.361069679260254, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.03535444661974907, + "learning_rate": 0.0005563037488593799, + "loss": 4.349765777587891, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03451581299304962, + "learning_rate": 0.0005561644577604074, + "loss": 4.343142509460449, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029810845851898193, + "learning_rate": 0.000556024962691404, + "loss": 4.307485580444336, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.027523217722773552, + "learning_rate": 0.0005558852637647584, + "loss": 4.366406440734863, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03043208085000515, + "learning_rate": 0.0005557453610930235, + "loss": 4.356929779052734, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.03058922104537487, + "learning_rate": 0.0005556052547889161, + "loss": 4.348942279815674, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.032052166759967804, + "learning_rate": 0.0005554649449653177, + "loss": 4.3496809005737305, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.03170694038271904, + "learning_rate": 0.0005553244317352731, + "loss": 4.315959930419922, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 4.4757890701293945, + "eval_runtime": 57.7451, + "eval_samples_per_second": 42.289, + "eval_steps_per_second": 1.333, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.029025450348854065, + "learning_rate": 0.0005551837152119915, + "loss": 4.288647174835205, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.029403962194919586, + "learning_rate": 0.0005550427955088455, + "loss": 4.288555145263672, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.03265019878745079, + "learning_rate": 0.0005549016727393715, + "loss": 4.292625904083252, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.0387202650308609, + "learning_rate": 0.0005547603470172697, + "loss": 4.238926887512207, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.04554206505417824, + "learning_rate": 0.0005546188184564036, + "loss": 4.29966402053833, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.036231476813554764, + "learning_rate": 0.0005544770871708004, + "loss": 4.304398536682129, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.038703009486198425, + "learning_rate": 0.0005543351532746501, + "loss": 4.2945356369018555, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03224619850516319, + "learning_rate": 0.0005541930168823067, + "loss": 4.261351585388184, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.029722856357693672, + "learning_rate": 0.0005540506781082864, + "loss": 4.303143501281738, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.02764045260846615, + "learning_rate": 0.0005539081370672695, + "loss": 4.243679523468018, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.0312200877815485, + "learning_rate": 0.0005537653938740984, + "loss": 4.279904365539551, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.027349065989255905, + "learning_rate": 0.0005536224486437788, + "loss": 4.270228385925293, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.030385233461856842, + "learning_rate": 0.0005534793014914793, + "loss": 4.301590442657471, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.03292274847626686, + "learning_rate": 0.0005533359525325307, + "loss": 4.301828384399414, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.03265977278351784, + "learning_rate": 0.0005531924018824266, + "loss": 4.286200523376465, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.03177836537361145, + "learning_rate": 0.0005530486496568236, + "loss": 4.296319007873535, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.028879934921860695, + "learning_rate": 0.0005529046959715397, + "loss": 4.261706829071045, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.02855812944471836, + "learning_rate": 0.0005527605409425562, + "loss": 4.273903846740723, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03037901036441326, + "learning_rate": 0.0005526161846860161, + "loss": 4.248080253601074, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.029279105365276337, + "learning_rate": 0.0005524716273182245, + "loss": 4.281311988830566, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.02592580020427704, + "learning_rate": 0.0005523268689556489, + "loss": 4.204848289489746, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.021899603307247162, + "learning_rate": 0.0005521819097149183, + "loss": 4.232453346252441, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.019704442471265793, + "learning_rate": 0.0005520367497128238, + "loss": 4.253462314605713, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.019550519064068794, + "learning_rate": 0.0005518913890663182, + "loss": 4.307737350463867, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.020320113748311996, + "learning_rate": 0.0005517458278925161, + "loss": 4.257411956787109, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02118559181690216, + "learning_rate": 0.0005516000663086932, + "loss": 4.262089729309082, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.02050616219639778, + "learning_rate": 0.0005514541044322872, + "loss": 4.2490644454956055, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019793255254626274, + "learning_rate": 0.0005513079423808967, + "loss": 4.267400741577148, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.020760254934430122, + "learning_rate": 0.0005511615802722823, + "loss": 4.2532429695129395, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.019524799659848213, + "learning_rate": 0.0005510150182243647, + "loss": 4.267422199249268, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.020166637375950813, + "learning_rate": 0.0005508682563552269, + "loss": 4.214296817779541, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.01964394748210907, + "learning_rate": 0.0005507212947831119, + "loss": 4.252087116241455, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01690433733165264, + "learning_rate": 0.0005505741336264239, + "loss": 4.232336044311523, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.017682762816548347, + "learning_rate": 0.0005504267730037283, + "loss": 4.213962554931641, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01688133366405964, + "learning_rate": 0.0005502792130337508, + "loss": 4.199461460113525, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.019166061654686928, + "learning_rate": 0.0005501314538353776, + "loss": 4.228936672210693, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.024065518751740456, + "learning_rate": 0.000549983495527656, + "loss": 4.214227676391602, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02518288791179657, + "learning_rate": 0.000549835338229793, + "loss": 4.256412029266357, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.026093967258930206, + "learning_rate": 0.0005496869820611563, + "loss": 4.216564178466797, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.023775635287165642, + "learning_rate": 0.0005495384271412739, + "loss": 4.226430892944336, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.018810907378792763, + "learning_rate": 0.0005493896735898338, + "loss": 4.190013885498047, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.018974872305989265, + "learning_rate": 0.000549240721526684, + "loss": 4.246231555938721, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.02069833129644394, + "learning_rate": 0.0005490915710718325, + "loss": 4.198457717895508, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.02097306028008461, + "learning_rate": 0.0005489422223454471, + "loss": 4.212507247924805, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022165542468428612, + "learning_rate": 0.0005487926754678556, + "loss": 4.2043843269348145, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.019068919122219086, + "learning_rate": 0.0005486429305595448, + "loss": 4.197460651397705, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01701812446117401, + "learning_rate": 0.0005484929877411618, + "loss": 4.220056533813477, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01629726029932499, + "learning_rate": 0.0005483428471335127, + "loss": 4.214868545532227, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018661806359887123, + "learning_rate": 0.000548192508857563, + "loss": 4.227047920227051, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.02005891129374504, + "learning_rate": 0.0005480419730344377, + "loss": 4.197265625, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.02591128647327423, + "learning_rate": 0.0005478912397854208, + "loss": 4.211041450500488, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.032659754157066345, + "learning_rate": 0.0005477403092319552, + "loss": 4.230195045471191, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.04057808965444565, + "learning_rate": 0.0005475891814956429, + "loss": 4.223843097686768, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.04103454574942589, + "learning_rate": 0.0005474378566982449, + "loss": 4.197145462036133, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.03810729831457138, + "learning_rate": 0.0005472863349616809, + "loss": 4.237957000732422, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.028838301077485085, + "learning_rate": 0.0005471346164080292, + "loss": 4.206722259521484, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02436191774904728, + "learning_rate": 0.0005469827011595266, + "loss": 4.1983137130737305, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.022839387878775597, + "learning_rate": 0.0005468305893385686, + "loss": 4.20530366897583, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.02375260926783085, + "learning_rate": 0.0005466782810677088, + "loss": 4.213202476501465, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.02177741937339306, + "learning_rate": 0.0005465257764696593, + "loss": 4.200539588928223, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.02351468615233898, + "learning_rate": 0.0005463730756672903, + "loss": 4.198091506958008, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.023375794291496277, + "learning_rate": 0.0005462201787836301, + "loss": 4.193244934082031, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.02479805052280426, + "learning_rate": 0.0005460670859418651, + "loss": 4.2250871658325195, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.026894841343164444, + "learning_rate": 0.000545913797265339, + "loss": 4.241791248321533, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.030582403764128685, + "learning_rate": 0.0005457603128775541, + "loss": 4.203666687011719, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.032222259789705276, + "learning_rate": 0.0005456066329021699, + "loss": 4.209629535675049, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.03181060031056404, + "learning_rate": 0.0005454527574630037, + "loss": 4.200459957122803, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02687576413154602, + "learning_rate": 0.0005452986866840298, + "loss": 4.211819648742676, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.025988668203353882, + "learning_rate": 0.0005451444206893806, + "loss": 4.198040962219238, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.02937682718038559, + "learning_rate": 0.0005449899596033453, + "loss": 4.222935676574707, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.03043951839208603, + "learning_rate": 0.0005448353035503704, + "loss": 4.2031755447387695, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.030401621013879776, + "learning_rate": 0.0005446804526550596, + "loss": 4.198380470275879, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 4.368444919586182, + "eval_runtime": 58.3009, + "eval_samples_per_second": 41.886, + "eval_steps_per_second": 1.321, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.03461585193872452, + "learning_rate": 0.0005445254070421732, + "loss": 4.147369384765625, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.039084456861019135, + "learning_rate": 0.0005443701668366288, + "loss": 4.124129295349121, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.04021848738193512, + "learning_rate": 0.0005442147321635007, + "loss": 4.127362251281738, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.041000768542289734, + "learning_rate": 0.00054405910314802, + "loss": 4.1432061195373535, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.04008613899350166, + "learning_rate": 0.0005439032799155737, + "loss": 4.169931888580322, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.039394307881593704, + "learning_rate": 0.000543747262591706, + "loss": 4.155813217163086, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.04374868795275688, + "learning_rate": 0.0005435910513021172, + "loss": 4.159891128540039, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0443582646548748, + "learning_rate": 0.0005434346461726641, + "loss": 4.157362937927246, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.04601646214723587, + "learning_rate": 0.0005432780473293593, + "loss": 4.152097225189209, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.04327242821455002, + "learning_rate": 0.0005431212548983718, + "loss": 4.170074462890625, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.0413462370634079, + "learning_rate": 0.0005429642690060262, + "loss": 4.156710147857666, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.036855220794677734, + "learning_rate": 0.0005428070897788034, + "loss": 4.138053894042969, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.03781217709183693, + "learning_rate": 0.0005426497173433398, + "loss": 4.172306060791016, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.03886301815509796, + "learning_rate": 0.0005424921518264275, + "loss": 4.135544776916504, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.03730994835495949, + "learning_rate": 0.0005423343933550143, + "loss": 4.171654224395752, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.03271950036287308, + "learning_rate": 0.0005421764420562032, + "loss": 4.169906139373779, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.03292673081159592, + "learning_rate": 0.0005420182980572527, + "loss": 4.163660049438477, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.029366906732320786, + "learning_rate": 0.0005418599614855768, + "loss": 4.161055088043213, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.023805437609553337, + "learning_rate": 0.0005417014324687442, + "loss": 4.172390937805176, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.023178137838840485, + "learning_rate": 0.000541542711134479, + "loss": 4.146788597106934, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.02403450198471546, + "learning_rate": 0.0005413837976106599, + "loss": 4.124364852905273, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.025009717792272568, + "learning_rate": 0.0005412246920253208, + "loss": 4.12275505065918, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.023431289941072464, + "learning_rate": 0.0005410653945066503, + "loss": 4.134739398956299, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.023803815245628357, + "learning_rate": 0.0005409059051829914, + "loss": 4.130463123321533, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.021219994872808456, + "learning_rate": 0.0005407462241828417, + "loss": 4.124577522277832, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.01899666152894497, + "learning_rate": 0.0005405863516348535, + "loss": 4.167086124420166, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.01889493502676487, + "learning_rate": 0.000540426287667833, + "loss": 4.152744293212891, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.019546914845705032, + "learning_rate": 0.000540266032410741, + "loss": 4.122655868530273, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02085448056459427, + "learning_rate": 0.0005401055859926923, + "loss": 4.141949653625488, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.019959403201937675, + "learning_rate": 0.0005399449485429555, + "loss": 4.131907939910889, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.01885392889380455, + "learning_rate": 0.0005397841201909535, + "loss": 4.118953227996826, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.018216131255030632, + "learning_rate": 0.0005396231010662627, + "loss": 4.096459865570068, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.019385207444429398, + "learning_rate": 0.0005394618912986132, + "loss": 4.111050128936768, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.018676603212952614, + "learning_rate": 0.0005393004910178889, + "loss": 4.156215667724609, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.01876339502632618, + "learning_rate": 0.0005391389003541271, + "loss": 4.1107282638549805, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.01831623539328575, + "learning_rate": 0.0005389771194375185, + "loss": 4.154116630554199, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.018896661698818207, + "learning_rate": 0.0005388151483984071, + "loss": 4.093320846557617, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.019582601264119148, + "learning_rate": 0.0005386529873672898, + "loss": 4.106500625610352, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.020313601940870285, + "learning_rate": 0.0005384906364748172, + "loss": 4.071961879730225, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.020385954529047012, + "learning_rate": 0.000538328095851792, + "loss": 4.129006862640381, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02231123298406601, + "learning_rate": 0.0005381653656291707, + "loss": 4.067756652832031, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.022957179695367813, + "learning_rate": 0.0005380024459380619, + "loss": 4.088230133056641, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021438462659716606, + "learning_rate": 0.000537839336909727, + "loss": 4.128036022186279, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01893841102719307, + "learning_rate": 0.00053767603867558, + "loss": 4.155407905578613, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.018539676442742348, + "learning_rate": 0.0005375125513671874, + "loss": 4.101880073547363, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.01881185919046402, + "learning_rate": 0.0005373488751162681, + "loss": 4.118906021118164, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.021657396107912064, + "learning_rate": 0.0005371850100546929, + "loss": 4.091935157775879, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.023293187841773033, + "learning_rate": 0.000537020956314485, + "loss": 4.105803489685059, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.021306317299604416, + "learning_rate": 0.0005368567140278196, + "loss": 4.0610809326171875, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.022078553214669228, + "learning_rate": 0.0005366922833270235, + "loss": 4.09867525100708, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.02266664244234562, + "learning_rate": 0.0005365276643445757, + "loss": 4.1027326583862305, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.02091757394373417, + "learning_rate": 0.0005363628572131069, + "loss": 4.080266952514648, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.020374659448862076, + "learning_rate": 0.0005361978620653991, + "loss": 4.088479042053223, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.02294750325381756, + "learning_rate": 0.0005360326790343858, + "loss": 4.084364891052246, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.02337353676557541, + "learning_rate": 0.000535867308253152, + "loss": 4.125696182250977, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.022643696516752243, + "learning_rate": 0.0005357017498549341, + "loss": 4.097160339355469, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.021080605685710907, + "learning_rate": 0.0005355360039731196, + "loss": 4.102363109588623, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.02269427664577961, + "learning_rate": 0.0005353700707412466, + "loss": 4.089844703674316, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.02195413038134575, + "learning_rate": 0.0005352039502930048, + "loss": 4.132293701171875, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.01805616170167923, + "learning_rate": 0.0005350376427622343, + "loss": 4.0842814445495605, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.018048783764243126, + "learning_rate": 0.0005348711482829261, + "loss": 4.109591007232666, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.0199508648365736, + "learning_rate": 0.000534704466989222, + "loss": 4.0908074378967285, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.020186785608530045, + "learning_rate": 0.0005345375990154138, + "loss": 4.077959060668945, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.022526193410158157, + "learning_rate": 0.0005343705444959441, + "loss": 4.109979629516602, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.025224441662430763, + "learning_rate": 0.000534203303565406, + "loss": 4.047533988952637, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.027292724698781967, + "learning_rate": 0.000534035876358542, + "loss": 4.09566593170166, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.025669090449810028, + "learning_rate": 0.0005338682630102454, + "loss": 4.0606160163879395, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02791150100529194, + "learning_rate": 0.0005337004636555595, + "loss": 4.097081184387207, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.03294133022427559, + "learning_rate": 0.0005335324784296768, + "loss": 4.121442794799805, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.027322066947817802, + "learning_rate": 0.0005333643074679403, + "loss": 4.1188740730285645, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.024866372346878052, + "learning_rate": 0.000533195950905842, + "loss": 4.082042694091797, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.02901027724146843, + "learning_rate": 0.0005330274088790239, + "loss": 4.110199451446533, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 4.283970355987549, + "eval_runtime": 58.4003, + "eval_samples_per_second": 41.815, + "eval_steps_per_second": 1.318, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.028958842158317566, + "learning_rate": 0.0005328586815232773, + "loss": 4.040733337402344, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.03491180017590523, + "learning_rate": 0.0005326897689745428, + "loss": 4.0473432540893555, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.04025818780064583, + "learning_rate": 0.0005325206713689101, + "loss": 4.021976470947266, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03998899459838867, + "learning_rate": 0.0005323513888426181, + "loss": 4.04901123046875, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.042165111750364304, + "learning_rate": 0.0005321819215320546, + "loss": 4.050230979919434, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.03511597216129303, + "learning_rate": 0.0005320122695737565, + "loss": 4.0989990234375, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.04064979776740074, + "learning_rate": 0.0005318424331044094, + "loss": 4.051448822021484, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.047536592930555344, + "learning_rate": 0.0005316724122608473, + "loss": 4.064278602600098, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.05592669919133186, + "learning_rate": 0.0005315022071800527, + "loss": 4.045516014099121, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.04262633994221687, + "learning_rate": 0.0005313318179991574, + "loss": 4.060666084289551, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03997781500220299, + "learning_rate": 0.0005311612448554402, + "loss": 4.065315246582031, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.038152433931827545, + "learning_rate": 0.0005309904878863293, + "loss": 4.069721221923828, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.0367623008787632, + "learning_rate": 0.0005308195472294001, + "loss": 4.080986976623535, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.0358598455786705, + "learning_rate": 0.0005306484230223769, + "loss": 4.065290927886963, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.039188165217638016, + "learning_rate": 0.000530477115403131, + "loss": 4.024333477020264, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.036839909851551056, + "learning_rate": 0.0005303056245096817, + "loss": 4.067532539367676, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.03456985577940941, + "learning_rate": 0.0005301339504801967, + "loss": 4.048443794250488, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.039487943053245544, + "learning_rate": 0.0005299620934529902, + "loss": 4.0585737228393555, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.03415827825665474, + "learning_rate": 0.0005297900535665244, + "loss": 4.06104850769043, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.03209620714187622, + "learning_rate": 0.0005296178309594087, + "loss": 4.009165287017822, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.028947360813617706, + "learning_rate": 0.0005294454257704, + "loss": 4.067384243011475, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026836825534701347, + "learning_rate": 0.0005292728381384018, + "loss": 4.046419620513916, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024095216765999794, + "learning_rate": 0.0005291000682024647, + "loss": 4.019008636474609, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.020965851843357086, + "learning_rate": 0.0005289271161017867, + "loss": 4.020613670349121, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.02250855043530464, + "learning_rate": 0.0005287539819757119, + "loss": 4.031634330749512, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.021385207772254944, + "learning_rate": 0.0005285806659637315, + "loss": 4.086246013641357, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.021104589104652405, + "learning_rate": 0.000528407168205483, + "loss": 4.033970355987549, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.019584037363529205, + "learning_rate": 0.0005282334888407503, + "loss": 4.025118827819824, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.02107258513569832, + "learning_rate": 0.000528059628009464, + "loss": 4.020823955535889, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.020438389852643013, + "learning_rate": 0.0005278855858517003, + "loss": 4.016323566436768, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.01881488971412182, + "learning_rate": 0.0005277113625076822, + "loss": 3.9991841316223145, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.01763618178665638, + "learning_rate": 0.000527536958117778, + "loss": 4.0062150955200195, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.017308862879872322, + "learning_rate": 0.0005273623728225022, + "loss": 3.946723461151123, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01706845872104168, + "learning_rate": 0.0005271876067625153, + "loss": 4.056562900543213, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.018664294853806496, + "learning_rate": 0.0005270126600786228, + "loss": 4.024652481079102, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.018523404374718666, + "learning_rate": 0.0005268375329117762, + "loss": 3.962064743041992, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.017408935353159904, + "learning_rate": 0.0005266622254030724, + "loss": 4.008697509765625, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.01588570512831211, + "learning_rate": 0.0005264867376937534, + "loss": 4.011664390563965, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.015084310434758663, + "learning_rate": 0.0005263110699252065, + "loss": 3.996224880218506, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01637054979801178, + "learning_rate": 0.000526135222238964, + "loss": 4.015972137451172, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.01673802360892296, + "learning_rate": 0.0005259591947767032, + "loss": 3.975414276123047, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.016544019803404808, + "learning_rate": 0.0005257829876802464, + "loss": 4.012955188751221, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.017455460503697395, + "learning_rate": 0.0005256066010915603, + "loss": 4.010180473327637, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.017625460401177406, + "learning_rate": 0.0005254300351527566, + "loss": 4.001679420471191, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.01732795126736164, + "learning_rate": 0.000525253290006091, + "loss": 4.006107330322266, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.016654646024107933, + "learning_rate": 0.0005250763657939641, + "loss": 4.045180797576904, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.015085658989846706, + "learning_rate": 0.0005248992626589203, + "loss": 3.994246482849121, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.015112568624317646, + "learning_rate": 0.0005247219807436489, + "loss": 4.004963397979736, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.014882124960422516, + "learning_rate": 0.000524544520190982, + "loss": 4.029565811157227, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.016012471169233322, + "learning_rate": 0.000524366881143897, + "loss": 4.018208026885986, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.015761924907565117, + "learning_rate": 0.0005241890637455141, + "loss": 3.9911205768585205, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.017770355567336082, + "learning_rate": 0.0005240110681390978, + "loss": 4.028733253479004, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01959751546382904, + "learning_rate": 0.0005238328944680558, + "loss": 4.003322124481201, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.018713824450969696, + "learning_rate": 0.0005236545428759394, + "loss": 3.991718292236328, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.01992208883166313, + "learning_rate": 0.0005234760135064434, + "loss": 4.002507209777832, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.021671291440725327, + "learning_rate": 0.0005232973065034055, + "loss": 4.021344184875488, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024282798171043396, + "learning_rate": 0.0005231184220108069, + "loss": 3.9899158477783203, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.023294344544410706, + "learning_rate": 0.0005229393601727713, + "loss": 4.046627044677734, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.02029707282781601, + "learning_rate": 0.000522760121133566, + "loss": 4.022454261779785, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.02151239849627018, + "learning_rate": 0.0005225807050376004, + "loss": 4.005476951599121, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.02320881187915802, + "learning_rate": 0.0005224011120294269, + "loss": 4.012726783752441, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.026484191417694092, + "learning_rate": 0.0005222213422537401, + "loss": 3.9904391765594482, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.026279661804437637, + "learning_rate": 0.0005220413958553777, + "loss": 3.9738848209381104, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.024299729615449905, + "learning_rate": 0.0005218612729793188, + "loss": 4.003316402435303, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.02337595447897911, + "learning_rate": 0.0005216809737706856, + "loss": 3.954984188079834, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.02416428178548813, + "learning_rate": 0.0005215004983747415, + "loss": 4.0317063331604, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.022488107904791832, + "learning_rate": 0.0005213198469368925, + "loss": 3.9736647605895996, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.019700225442647934, + "learning_rate": 0.0005211390196026861, + "loss": 3.979867458343506, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.020382333546876907, + "learning_rate": 0.0005209580165178117, + "loss": 3.995251178741455, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.020149298012256622, + "learning_rate": 0.0005207768378281003, + "loss": 3.9987268447875977, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.019790509715676308, + "learning_rate": 0.0005205954836795241, + "loss": 3.9838991165161133, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.022096451371908188, + "learning_rate": 0.000520413954218197, + "loss": 3.995800256729126, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.214347839355469, + "eval_runtime": 58.8872, + "eval_samples_per_second": 41.469, + "eval_steps_per_second": 1.308, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02203402668237686, + "learning_rate": 0.0005202322495903739, + "loss": 3.915332078933716, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.02336675301194191, + "learning_rate": 0.0005200503699424509, + "loss": 3.914801597595215, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.027339637279510498, + "learning_rate": 0.0005198683154209654, + "loss": 3.9277780055999756, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.026872653514146805, + "learning_rate": 0.0005196860861725952, + "loss": 3.9160094261169434, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.026558255776762962, + "learning_rate": 0.0005195036823441592, + "loss": 3.9084300994873047, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.02644367143511772, + "learning_rate": 0.0005193211040826169, + "loss": 3.9328718185424805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.024677276611328125, + "learning_rate": 0.0005191383515350681, + "loss": 3.88993501663208, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.023194244131445885, + "learning_rate": 0.0005189554248487536, + "loss": 3.923567056655884, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.024653147906064987, + "learning_rate": 0.0005187723241710539, + "loss": 3.8740999698638916, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.026041202247142792, + "learning_rate": 0.0005185890496494899, + "loss": 3.9542148113250732, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.029586758464574814, + "learning_rate": 0.0005184056014317225, + "loss": 3.9241347312927246, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.0325787179172039, + "learning_rate": 0.0005182219796655528, + "loss": 3.9656593799591064, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.03183380514383316, + "learning_rate": 0.0005180381844989214, + "loss": 3.950648307800293, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.03434471786022186, + "learning_rate": 0.0005178542160799088, + "loss": 3.9590439796447754, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.030099432915449142, + "learning_rate": 0.0005176700745567353, + "loss": 3.9518113136291504, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.025950564071536064, + "learning_rate": 0.0005174857600777599, + "loss": 3.8908939361572266, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.029959935694932938, + "learning_rate": 0.0005173012727914816, + "loss": 3.945037364959717, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.029551763087511063, + "learning_rate": 0.0005171166128465389, + "loss": 3.923959732055664, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.03538523241877556, + "learning_rate": 0.0005169317803917083, + "loss": 3.9395527839660645, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.037924155592918396, + "learning_rate": 0.0005167467755759065, + "loss": 3.945115089416504, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.036773014813661575, + "learning_rate": 0.0005165615985481881, + "loss": 3.9484646320343018, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.04125838354229927, + "learning_rate": 0.0005163762494577474, + "loss": 4.000699043273926, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.044125135987997055, + "learning_rate": 0.0005161907284539165, + "loss": 3.9828028678894043, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.040853384882211685, + "learning_rate": 0.0005160050356861663, + "loss": 3.957505702972412, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.03501008450984955, + "learning_rate": 0.0005158191713041063, + "loss": 3.9360849857330322, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.04082556813955307, + "learning_rate": 0.0005156331354574838, + "loss": 3.95941424369812, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.045432962477207184, + "learning_rate": 0.0005154469282961847, + "loss": 3.9666922092437744, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.04048244655132294, + "learning_rate": 0.0005152605499702328, + "loss": 3.984555244445801, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.031016314402222633, + "learning_rate": 0.0005150740006297896, + "loss": 3.945204257965088, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.03210771828889847, + "learning_rate": 0.0005148872804251548, + "loss": 3.976186752319336, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.03130023553967476, + "learning_rate": 0.000514700389506765, + "loss": 3.915727138519287, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.031193193048238754, + "learning_rate": 0.0005145133280251954, + "loss": 3.961866855621338, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.027518954128026962, + "learning_rate": 0.0005143260961311578, + "loss": 3.9742956161499023, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.02751155197620392, + "learning_rate": 0.0005141386939755014, + "loss": 3.949662208557129, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.026033353060483932, + "learning_rate": 0.0005139511217092129, + "loss": 3.9327926635742188, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.02579621411859989, + "learning_rate": 0.000513763379483416, + "loss": 3.9334418773651123, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.023009559139609337, + "learning_rate": 0.0005135754674493709, + "loss": 3.969707489013672, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.02196447364985943, + "learning_rate": 0.0005133873857584753, + "loss": 3.9084434509277344, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.02206484042108059, + "learning_rate": 0.0005131991345622629, + "loss": 3.922117233276367, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.023360205814242363, + "learning_rate": 0.0005130107140124046, + "loss": 3.965158224105835, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02162669226527214, + "learning_rate": 0.0005128221242607072, + "loss": 3.9519729614257812, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.019719846546649933, + "learning_rate": 0.0005126333654591141, + "loss": 3.9195961952209473, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.018988260999321938, + "learning_rate": 0.0005124444377597049, + "loss": 3.937926769256592, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.019278401508927345, + "learning_rate": 0.0005122553413146951, + "loss": 3.9126808643341064, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019101005047559738, + "learning_rate": 0.0005120660762764366, + "loss": 3.8993749618530273, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.019156700000166893, + "learning_rate": 0.0005118766427974166, + "loss": 3.9531068801879883, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019944407045841217, + "learning_rate": 0.0005116870410302582, + "loss": 3.9645180702209473, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.01826174184679985, + "learning_rate": 0.0005114972711277202, + "loss": 3.948477268218994, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.0191368218511343, + "learning_rate": 0.0005113073332426968, + "loss": 3.944648027420044, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.017768343910574913, + "learning_rate": 0.0005111172275282174, + "loss": 3.9584474563598633, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.01955590397119522, + "learning_rate": 0.0005109269541374469, + "loss": 3.8936800956726074, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.02100536972284317, + "learning_rate": 0.000510736513223685, + "loss": 3.93919038772583, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.020873935893177986, + "learning_rate": 0.0005105459049403665, + "loss": 3.929244041442871, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.021876046434044838, + "learning_rate": 0.0005103551294410614, + "loss": 3.9264731407165527, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.021726684644818306, + "learning_rate": 0.0005101641868794736, + "loss": 3.923135280609131, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.019308188930153847, + "learning_rate": 0.0005099730774094423, + "loss": 3.9062135219573975, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01917017623782158, + "learning_rate": 0.0005097818011849408, + "loss": 3.898810386657715, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017933931201696396, + "learning_rate": 0.0005095903583600772, + "loss": 3.8899734020233154, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.015131795778870583, + "learning_rate": 0.0005093987490890933, + "loss": 3.9452900886535645, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.017190825194120407, + "learning_rate": 0.0005092069735263651, + "loss": 3.9124746322631836, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.01618487387895584, + "learning_rate": 0.000509015031826403, + "loss": 3.91654109954834, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.014920198358595371, + "learning_rate": 0.0005088229241438509, + "loss": 3.9144535064697266, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01615205965936184, + "learning_rate": 0.0005086306506334864, + "loss": 3.917003631591797, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.015370192006230354, + "learning_rate": 0.0005084382114502208, + "loss": 3.8910603523254395, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.015290996059775352, + "learning_rate": 0.0005082456067490988, + "loss": 3.9122695922851562, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.015913106501102448, + "learning_rate": 0.0005080528366852987, + "loss": 3.9304556846618652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.01639600843191147, + "learning_rate": 0.0005078599014141318, + "loss": 3.8974502086639404, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.017010601237416267, + "learning_rate": 0.0005076668010910425, + "loss": 3.9723148345947266, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.017778996378183365, + "learning_rate": 0.0005074735358716083, + "loss": 3.8905391693115234, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.020029611885547638, + "learning_rate": 0.0005072801059115394, + "loss": 3.934976100921631, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.01931857503950596, + "learning_rate": 0.000507086511366679, + "loss": 3.9152774810791016, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019750485196709633, + "learning_rate": 0.0005068927523930027, + "loss": 3.877650499343872, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.161914825439453, + "eval_runtime": 58.5217, + "eval_samples_per_second": 41.728, + "eval_steps_per_second": 1.316, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02073049545288086, + "learning_rate": 0.0005066988291466184, + "loss": 3.8331246376037598, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02272389829158783, + "learning_rate": 0.0005065047417837669, + "loss": 3.8847155570983887, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02139078453183174, + "learning_rate": 0.0005063104904608206, + "loss": 3.8276169300079346, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02043229155242443, + "learning_rate": 0.0005061160753342843, + "loss": 3.8804001808166504, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.02345564216375351, + "learning_rate": 0.0005059214965607948, + "loss": 3.851609468460083, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.023448185995221138, + "learning_rate": 0.0005057267542971207, + "loss": 3.8446507453918457, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.024826066568493843, + "learning_rate": 0.0005055318487001624, + "loss": 3.86712384223938, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.02491980977356434, + "learning_rate": 0.0005053367799269518, + "loss": 3.8374528884887695, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.023002834990620613, + "learning_rate": 0.0005051415481346522, + "loss": 3.783214569091797, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.02123952843248844, + "learning_rate": 0.0005049461534805585, + "loss": 3.8463428020477295, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.01981225423514843, + "learning_rate": 0.0005047505961220964, + "loss": 3.8307082653045654, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01915367692708969, + "learning_rate": 0.0005045548762168231, + "loss": 3.8190226554870605, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.01890859752893448, + "learning_rate": 0.0005043589939224266, + "loss": 3.8428220748901367, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.019150376319885254, + "learning_rate": 0.0005041629493967257, + "loss": 3.830247402191162, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.020099453628063202, + "learning_rate": 0.0005039667427976697, + "loss": 3.838806629180908, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.021136781200766563, + "learning_rate": 0.0005037703742833391, + "loss": 3.8344273567199707, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.022857150062918663, + "learning_rate": 0.0005035738440119442, + "loss": 3.8317880630493164, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.024891739711165428, + "learning_rate": 0.0005033771521418261, + "loss": 3.837855577468872, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.02595464698970318, + "learning_rate": 0.0005031802988314557, + "loss": 3.846050262451172, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.02722409926354885, + "learning_rate": 0.0005029832842394342, + "loss": 3.8164305686950684, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.024486860260367393, + "learning_rate": 0.0005027861085244927, + "loss": 3.8099875450134277, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.021551169455051422, + "learning_rate": 0.0005025887718454922, + "loss": 3.825925827026367, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.02121506817638874, + "learning_rate": 0.0005023912743614232, + "loss": 3.870727777481079, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.01948230341076851, + "learning_rate": 0.0005021936162314058, + "loss": 3.8035314083099365, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.01896044984459877, + "learning_rate": 0.0005019957976146898, + "loss": 3.84665584564209, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.020620400086045265, + "learning_rate": 0.0005017978186706538, + "loss": 3.8338310718536377, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.020855262875556946, + "learning_rate": 0.0005015996795588062, + "loss": 3.8781003952026367, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.02153870463371277, + "learning_rate": 0.0005014013804387838, + "loss": 3.8783490657806396, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.023751024156808853, + "learning_rate": 0.0005012029214703529, + "loss": 3.846198558807373, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.023987066000699997, + "learning_rate": 0.000501004302813408, + "loss": 3.834296226501465, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.023626182228326797, + "learning_rate": 0.0005008055246279727, + "loss": 3.8591227531433105, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.025707192718982697, + "learning_rate": 0.000500606587074199, + "loss": 3.8766489028930664, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.027723029255867004, + "learning_rate": 0.0005004074903123674, + "loss": 3.88264799118042, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.0278895553201437, + "learning_rate": 0.0005002082345028864, + "loss": 3.8932995796203613, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.025622662156820297, + "learning_rate": 0.0005000088198062929, + "loss": 3.8734612464904785, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.02639615908265114, + "learning_rate": 0.0004998092463832516, + "loss": 3.786813735961914, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.026765599846839905, + "learning_rate": 0.0004996095143945554, + "loss": 3.853966474533081, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02823319099843502, + "learning_rate": 0.0004994096240011246, + "loss": 3.8321340084075928, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.023394955322146416, + "learning_rate": 0.0004992095753640072, + "loss": 3.8527331352233887, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.023874662816524506, + "learning_rate": 0.0004990093686443791, + "loss": 3.8822879791259766, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.02301311306655407, + "learning_rate": 0.000498809004003543, + "loss": 3.846186876296997, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.022630592808127403, + "learning_rate": 0.0004986084816029291, + "loss": 3.867992639541626, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.02424517087638378, + "learning_rate": 0.0004984078016040946, + "loss": 3.867293357849121, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.024116624146699905, + "learning_rate": 0.000498206964168724, + "loss": 3.846282958984375, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.02187768742442131, + "learning_rate": 0.000498005969458628, + "loss": 3.8429484367370605, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.01879809983074665, + "learning_rate": 0.0004978048176357447, + "loss": 3.863215446472168, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.020426234230399132, + "learning_rate": 0.0004976035088621385, + "loss": 3.825230121612549, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01992812193930149, + "learning_rate": 0.0004974020433000003, + "loss": 3.851665496826172, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.020693328231573105, + "learning_rate": 0.000497200421111647, + "loss": 3.906999111175537, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.021137919276952744, + "learning_rate": 0.0004969986424595221, + "loss": 3.8369035720825195, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.023285407572984695, + "learning_rate": 0.0004967967075061949, + "loss": 3.8658390045166016, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.0219135619699955, + "learning_rate": 0.0004965946164143609, + "loss": 3.873072862625122, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.020101983100175858, + "learning_rate": 0.0004963923693468413, + "loss": 3.847805976867676, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.01875176839530468, + "learning_rate": 0.0004961899664665827, + "loss": 3.7976622581481934, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01913360133767128, + "learning_rate": 0.0004959874079366576, + "loss": 3.853851318359375, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.016863619908690453, + "learning_rate": 0.0004957846939202637, + "loss": 3.877378225326538, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.017802366986870766, + "learning_rate": 0.000495581824580724, + "loss": 3.861356019973755, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.01981814205646515, + "learning_rate": 0.0004953788000814866, + "loss": 3.88613224029541, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.017508897930383682, + "learning_rate": 0.000495175620586125, + "loss": 3.8202030658721924, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.019240548834204674, + "learning_rate": 0.0004949722862583368, + "loss": 3.853827714920044, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.022305598482489586, + "learning_rate": 0.000494768797261945, + "loss": 3.885148763656616, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.02531025931239128, + "learning_rate": 0.000494565153760897, + "loss": 3.8412041664123535, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.024474849924445152, + "learning_rate": 0.0004943613559192648, + "loss": 3.865644931793213, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.020430924370884895, + "learning_rate": 0.0004941574039012444, + "loss": 3.8649442195892334, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.01906431093811989, + "learning_rate": 0.0004939532978711566, + "loss": 3.8462204933166504, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.021118057891726494, + "learning_rate": 0.0004937490379934456, + "loss": 3.8652522563934326, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.020337693393230438, + "learning_rate": 0.00049354462443268, + "loss": 3.8506221771240234, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.019090797752141953, + "learning_rate": 0.0004933400573535521, + "loss": 3.881237506866455, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.020973797887563705, + "learning_rate": 0.000493135336920878, + "loss": 3.821274757385254, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.021984362974762917, + "learning_rate": 0.0004929304632995973, + "loss": 3.8793835639953613, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.0215684212744236, + "learning_rate": 0.0004927254366547727, + "loss": 3.8259663581848145, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.021673865616321564, + "learning_rate": 0.0004925202571515908, + "loss": 3.8505711555480957, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.123871326446533, + "eval_runtime": 58.8357, + "eval_samples_per_second": 41.505, + "eval_steps_per_second": 1.309, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.0203359667211771, + "learning_rate": 0.0004923149249553608, + "loss": 3.785353660583496, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.023197868838906288, + "learning_rate": 0.0004921094402315152, + "loss": 3.7479515075683594, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.02314298041164875, + "learning_rate": 0.0004919038031456093, + "loss": 3.7697315216064453, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.0256908368319273, + "learning_rate": 0.0004916980138633213, + "loss": 3.79056978225708, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.026193121448159218, + "learning_rate": 0.0004914920725504518, + "loss": 3.7722532749176025, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.027120303362607956, + "learning_rate": 0.000491285979372924, + "loss": 3.7991981506347656, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.0266975536942482, + "learning_rate": 0.0004910797344967834, + "loss": 3.778301477432251, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.02806396596133709, + "learning_rate": 0.000490873338088198, + "loss": 3.7578704357147217, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.02893839031457901, + "learning_rate": 0.0004906667903134573, + "loss": 3.7401275634765625, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.031170370057225227, + "learning_rate": 0.0004904600913389735, + "loss": 3.780686616897583, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.03364266827702522, + "learning_rate": 0.0004902532413312799, + "loss": 3.794442653656006, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.030121006071567535, + "learning_rate": 0.0004900462404570321, + "loss": 3.779569149017334, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.03296204283833504, + "learning_rate": 0.0004898390888830067, + "loss": 3.7960336208343506, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.03381749242544174, + "learning_rate": 0.0004896317867761022, + "loss": 3.7959280014038086, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.035383883863687515, + "learning_rate": 0.000489424334303338, + "loss": 3.8008625507354736, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.03759761154651642, + "learning_rate": 0.0004892167316318548, + "loss": 3.7997028827667236, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.030651796609163284, + "learning_rate": 0.0004890089789289147, + "loss": 3.81191349029541, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.0262477844953537, + "learning_rate": 0.0004888010763618997, + "loss": 3.8046622276306152, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.030398324131965637, + "learning_rate": 0.0004885930240983136, + "loss": 3.7903637886047363, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.030515259131789207, + "learning_rate": 0.0004883848223057801, + "loss": 3.8111231327056885, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.031686048954725266, + "learning_rate": 0.00048817647115204386, + "loss": 3.781445026397705, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.025968654081225395, + "learning_rate": 0.00048796797080496936, + "loss": 3.788196563720703, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.025340190157294273, + "learning_rate": 0.00048775932143254176, + "loss": 3.763331413269043, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.024738596752285957, + "learning_rate": 0.0004875505232028661, + "loss": 3.8005969524383545, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.021990075707435608, + "learning_rate": 0.0004873415762841672, + "loss": 3.795844078063965, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.02191784791648388, + "learning_rate": 0.00048713248084478986, + "loss": 3.8062384128570557, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.023678259924054146, + "learning_rate": 0.0004869232370531986, + "loss": 3.7760679721832275, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.024137988686561584, + "learning_rate": 0.0004867138450779773, + "loss": 3.7917492389678955, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.02673223614692688, + "learning_rate": 0.0004865043050878293, + "loss": 3.8006649017333984, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.02395169623196125, + "learning_rate": 0.0004862946172515771, + "loss": 3.7902212142944336, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.02410844899713993, + "learning_rate": 0.0004860847817381625, + "loss": 3.8073291778564453, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.028109831735491753, + "learning_rate": 0.0004858747987166463, + "loss": 3.80116605758667, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.025538964197039604, + "learning_rate": 0.00048566466835620774, + "loss": 3.791151523590088, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.02393815852701664, + "learning_rate": 0.0004854543908261453, + "loss": 3.7766106128692627, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.024659741669893265, + "learning_rate": 0.0004852439662958756, + "loss": 3.789135217666626, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.02668776921927929, + "learning_rate": 0.000485033394934934, + "loss": 3.797905445098877, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.0258641317486763, + "learning_rate": 0.000484822676912974, + "loss": 3.783323287963867, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.024846022948622704, + "learning_rate": 0.0004846118123997674, + "loss": 3.7976717948913574, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.024981874972581863, + "learning_rate": 0.00048440080156520387, + "loss": 3.818732738494873, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.02556881681084633, + "learning_rate": 0.0004841896445792909, + "loss": 3.7821896076202393, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.02558571845293045, + "learning_rate": 0.000483978341612154, + "loss": 3.8143680095672607, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.02361566200852394, + "learning_rate": 0.0004837668928340362, + "loss": 3.7710189819335938, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022570716217160225, + "learning_rate": 0.0004835552984152978, + "loss": 3.775564670562744, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.022478004917502403, + "learning_rate": 0.00048334355852641664, + "loss": 3.7920689582824707, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021241728216409683, + "learning_rate": 0.0004831316733379879, + "loss": 3.7857837677001953, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.018921511247754097, + "learning_rate": 0.0004829196430207235, + "loss": 3.7659268379211426, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01926037296652794, + "learning_rate": 0.00048270746774545237, + "loss": 3.7527129650115967, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.018743667751550674, + "learning_rate": 0.0004824951476831204, + "loss": 3.767991065979004, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.017273949459195137, + "learning_rate": 0.00048228268300479015, + "loss": 3.7834174633026123, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.017054244875907898, + "learning_rate": 0.0004820700738816403, + "loss": 3.7775752544403076, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017190931364893913, + "learning_rate": 0.0004818573204849664, + "loss": 3.7709648609161377, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01665034517645836, + "learning_rate": 0.00048164442298617987, + "loss": 3.7575321197509766, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.016932843253016472, + "learning_rate": 0.00048143138155680844, + "loss": 3.7922799587249756, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01702750287950039, + "learning_rate": 0.0004812181963684958, + "loss": 3.7729780673980713, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01766599342226982, + "learning_rate": 0.00048100486759300134, + "loss": 3.7876572608947754, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01785697042942047, + "learning_rate": 0.00048079139540220024, + "loss": 3.8131699562072754, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01830688677728176, + "learning_rate": 0.0004805777799680831, + "loss": 3.790778636932373, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017348438501358032, + "learning_rate": 0.0004803640214627561, + "loss": 3.7984330654144287, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017999377101659775, + "learning_rate": 0.0004801501200584406, + "loss": 3.7756147384643555, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.01801210641860962, + "learning_rate": 0.000479936075927473, + "loss": 3.789759874343872, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.018445398658514023, + "learning_rate": 0.000479721889242305, + "loss": 3.7683000564575195, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.017544370144605637, + "learning_rate": 0.0004795075601755027, + "loss": 3.7776038646698, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.019834233447909355, + "learning_rate": 0.00047929308889974737, + "loss": 3.7559571266174316, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01791626401245594, + "learning_rate": 0.0004790784755878346, + "loss": 3.7746729850769043, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.017638562247157097, + "learning_rate": 0.00047886372041267434, + "loss": 3.7933082580566406, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.018974050879478455, + "learning_rate": 0.0004786488235472909, + "loss": 3.795713424682617, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.018694303929805756, + "learning_rate": 0.0004784337851648231, + "loss": 3.7675955295562744, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.01788092777132988, + "learning_rate": 0.0004782186054385233, + "loss": 3.786083698272705, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.01902497000992298, + "learning_rate": 0.0004780032845417579, + "loss": 3.7976233959198, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.020558880642056465, + "learning_rate": 0.000477787822648007, + "loss": 3.7500152587890625, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.020093899220228195, + "learning_rate": 0.00047757221993086443, + "loss": 3.703448534011841, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.019833799451589584, + "learning_rate": 0.00047735647656403754, + "loss": 3.756059408187866, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.093987941741943, + "eval_runtime": 57.6894, + "eval_samples_per_second": 42.33, + "eval_steps_per_second": 1.335, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.022167526185512543, + "learning_rate": 0.0004771405927213467, + "loss": 3.696444511413574, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.027874166145920753, + "learning_rate": 0.0004769245685767255, + "loss": 3.713630199432373, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.030872784554958344, + "learning_rate": 0.00047670840430422113, + "loss": 3.678295612335205, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.03246859833598137, + "learning_rate": 0.0004764921000779928, + "loss": 3.7186388969421387, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.03667069226503372, + "learning_rate": 0.0004762756560723131, + "loss": 3.7457995414733887, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.036286015063524246, + "learning_rate": 0.0004760590724615671, + "loss": 3.6881775856018066, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.03413614630699158, + "learning_rate": 0.0004758423494202522, + "loss": 3.6999263763427734, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.03384552150964737, + "learning_rate": 0.0004756254871229782, + "loss": 3.6959803104400635, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.03227181360125542, + "learning_rate": 0.00047540848574446724, + "loss": 3.7182722091674805, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.036366235464811325, + "learning_rate": 0.0004751913454595535, + "loss": 3.7116851806640625, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.03573326766490936, + "learning_rate": 0.0004749740664431828, + "loss": 3.741865873336792, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.03482096269726753, + "learning_rate": 0.000474756648870413, + "loss": 3.7195520401000977, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.03232625126838684, + "learning_rate": 0.0004745390929164134, + "loss": 3.7443251609802246, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.03421664610505104, + "learning_rate": 0.0004743213987564651, + "loss": 3.6867361068725586, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.03585013002157211, + "learning_rate": 0.0004741035665659603, + "loss": 3.7492403984069824, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.029985329136252403, + "learning_rate": 0.00047388559652040236, + "loss": 3.705209493637085, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.02729589305818081, + "learning_rate": 0.0004736674887954059, + "loss": 3.711106538772583, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.029642216861248016, + "learning_rate": 0.0004734492435666963, + "loss": 3.7274727821350098, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.02816903032362461, + "learning_rate": 0.0004732308610101099, + "loss": 3.7181742191314697, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.027143625542521477, + "learning_rate": 0.00047301234130159346, + "loss": 3.719235897064209, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.025802744552493095, + "learning_rate": 0.00047279368461720445, + "loss": 3.771820545196533, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.026424532756209373, + "learning_rate": 0.0004725748911331106, + "loss": 3.7516651153564453, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.029677648097276688, + "learning_rate": 0.0004723559610255898, + "loss": 3.7058472633361816, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.031548649072647095, + "learning_rate": 0.0004721368944710302, + "loss": 3.756992816925049, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.027667395770549774, + "learning_rate": 0.00047191769164592974, + "loss": 3.742711067199707, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.027639545500278473, + "learning_rate": 0.000471698352726896, + "loss": 3.728883743286133, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.02950332500040531, + "learning_rate": 0.0004714788778906467, + "loss": 3.7160212993621826, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.026302607730031013, + "learning_rate": 0.0004712592673140084, + "loss": 3.735713481903076, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.022483112290501595, + "learning_rate": 0.00047103952117391764, + "loss": 3.702686309814453, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.021729422733187675, + "learning_rate": 0.00047081963964741986, + "loss": 3.7195816040039062, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.021159254014492035, + "learning_rate": 0.0004705996229116696, + "loss": 3.684321880340576, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.02233424223959446, + "learning_rate": 0.0004703794711439304, + "loss": 3.7233057022094727, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.02142048068344593, + "learning_rate": 0.0004701591845215744, + "loss": 3.7083680629730225, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.022331461310386658, + "learning_rate": 0.00046993876322208274, + "loss": 3.7408461570739746, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.020629866048693657, + "learning_rate": 0.0004697182074230448, + "loss": 3.759129047393799, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.02174685336649418, + "learning_rate": 0.0004694975173021586, + "loss": 3.72351336479187, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.021538110449910164, + "learning_rate": 0.0004692766930372299, + "loss": 3.716416597366333, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.020800625905394554, + "learning_rate": 0.00046905573480617285, + "loss": 3.726954936981201, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.02039588987827301, + "learning_rate": 0.00046883464278700966, + "loss": 3.7336390018463135, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.02107219770550728, + "learning_rate": 0.0004686134171578702, + "loss": 3.722311496734619, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.02068956382572651, + "learning_rate": 0.00046839205809699183, + "loss": 3.767139434814453, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.020190054550766945, + "learning_rate": 0.0004681705657827195, + "loss": 3.6960511207580566, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.018814411014318466, + "learning_rate": 0.0004679489403935058, + "loss": 3.70375919342041, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.017568519338965416, + "learning_rate": 0.00046772718210791005, + "loss": 3.756627321243286, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.018188195303082466, + "learning_rate": 0.00046750529110459903, + "loss": 3.6964683532714844, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.018925746902823448, + "learning_rate": 0.0004672832675623463, + "loss": 3.731172561645508, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.01690756157040596, + "learning_rate": 0.0004670611116600322, + "loss": 3.6955318450927734, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016300756484270096, + "learning_rate": 0.00046683882357664375, + "loss": 3.762632369995117, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.016405068337917328, + "learning_rate": 0.0004666164034912744, + "loss": 3.75130558013916, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.017687877640128136, + "learning_rate": 0.00046639385158312406, + "loss": 3.687530755996704, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.017524344846606255, + "learning_rate": 0.00046617116803149874, + "loss": 3.738790273666382, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.01714804768562317, + "learning_rate": 0.00046594835301581053, + "loss": 3.69827938079834, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018042324110865593, + "learning_rate": 0.00046572540671557745, + "loss": 3.71600341796875, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018648479133844376, + "learning_rate": 0.00046550232931042346, + "loss": 3.742591619491577, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.019198937341570854, + "learning_rate": 0.00046527912098007787, + "loss": 3.739820957183838, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.01796373911201954, + "learning_rate": 0.00046505578190437564, + "loss": 3.719025135040283, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01836620457470417, + "learning_rate": 0.0004648323122632569, + "loss": 3.741434097290039, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.02192627638578415, + "learning_rate": 0.0004646087122367673, + "loss": 3.7553882598876953, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.02147185243666172, + "learning_rate": 0.00046438498200505727, + "loss": 3.7705790996551514, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01855960488319397, + "learning_rate": 0.0004641611217483821, + "loss": 3.704925537109375, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.01925475336611271, + "learning_rate": 0.00046393713164710217, + "loss": 3.711491346359253, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019796503707766533, + "learning_rate": 0.00046371301188168204, + "loss": 3.7241992950439453, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.018511971458792686, + "learning_rate": 0.0004634887626326911, + "loss": 3.6785435676574707, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.018214019015431404, + "learning_rate": 0.00046326438408080293, + "loss": 3.7299962043762207, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.018912190571427345, + "learning_rate": 0.00046303987640679517, + "loss": 3.773698329925537, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.01939854398369789, + "learning_rate": 0.0004628152397915498, + "loss": 3.682321071624756, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.017859861254692078, + "learning_rate": 0.00046259047441605215, + "loss": 3.7254951000213623, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.019249707460403442, + "learning_rate": 0.0004623655804613919, + "loss": 3.7263433933258057, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.019221967086195946, + "learning_rate": 0.00046214055810876194, + "loss": 3.7229225635528564, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01871953159570694, + "learning_rate": 0.00046191540753945886, + "loss": 3.7248311042785645, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01747182011604309, + "learning_rate": 0.00046169012893488204, + "loss": 3.699913740158081, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.019630730152130127, + "learning_rate": 0.00046146472247653484, + "loss": 3.787783145904541, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.072091102600098, + "eval_runtime": 63.1589, + "eval_samples_per_second": 38.664, + "eval_steps_per_second": 1.219, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.022918934002518654, + "learning_rate": 0.0004612391883460228, + "loss": 3.666675567626953, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.0264838095754385, + "learning_rate": 0.00046101352672505493, + "loss": 3.6371259689331055, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.028463594615459442, + "learning_rate": 0.0004607877377954426, + "loss": 3.657130241394043, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.02900792844593525, + "learning_rate": 0.00046056182173909995, + "loss": 3.6365628242492676, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.028593607246875763, + "learning_rate": 0.0004603357787380435, + "loss": 3.647258996963501, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.02878790907561779, + "learning_rate": 0.0004601096089743919, + "loss": 3.682035446166992, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.02980183996260166, + "learning_rate": 0.0004598833126303661, + "loss": 3.623297691345215, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.03375725448131561, + "learning_rate": 0.00045965688988828884, + "loss": 3.6240830421447754, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.035045325756073, + "learning_rate": 0.00045943034093058507, + "loss": 3.6372249126434326, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.03519923985004425, + "learning_rate": 0.000459203665939781, + "loss": 3.6157641410827637, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.03468473628163338, + "learning_rate": 0.00045897686509850487, + "loss": 3.6553335189819336, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.034817393869161606, + "learning_rate": 0.0004587499385894856, + "loss": 3.653806686401367, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.03549059107899666, + "learning_rate": 0.000458522886595554, + "loss": 3.6730189323425293, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.028001410886645317, + "learning_rate": 0.0004582957092996418, + "loss": 3.662490129470825, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.02532520331442356, + "learning_rate": 0.0004580684068847817, + "loss": 3.629936456680298, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.028215918689966202, + "learning_rate": 0.0004578409795341069, + "loss": 3.6507444381713867, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.026823483407497406, + "learning_rate": 0.0004576134274308517, + "loss": 3.661811113357544, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.0281058382242918, + "learning_rate": 0.0004573857507583506, + "loss": 3.6805872917175293, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.02548503689467907, + "learning_rate": 0.00045715794970003874, + "loss": 3.6586413383483887, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.024795135483145714, + "learning_rate": 0.0004569300244394513, + "loss": 3.654989004135132, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.025501688942313194, + "learning_rate": 0.0004567019751602234, + "loss": 3.6915903091430664, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.02377370186150074, + "learning_rate": 0.00045647380204609034, + "loss": 3.660588264465332, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.022150425240397453, + "learning_rate": 0.00045624550528088704, + "loss": 3.64870023727417, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.02415974996984005, + "learning_rate": 0.00045601708504854803, + "loss": 3.6498191356658936, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.02140078693628311, + "learning_rate": 0.0004557885415331074, + "loss": 3.703489065170288, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.022121304646134377, + "learning_rate": 0.00045555987491869865, + "loss": 3.6606979370117188, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.02168990671634674, + "learning_rate": 0.00045533108538955416, + "loss": 3.6801178455352783, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.020228713750839233, + "learning_rate": 0.0004551021731300055, + "loss": 3.6707205772399902, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.018500352278351784, + "learning_rate": 0.0004548731383244834, + "loss": 3.647545576095581, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.020223816856741905, + "learning_rate": 0.0004546439811575169, + "loss": 3.644932985305786, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.019502121955156326, + "learning_rate": 0.00045441470181373375, + "loss": 3.62888240814209, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.01926770992577076, + "learning_rate": 0.0004541853004778603, + "loss": 3.675457715988159, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01888471283018589, + "learning_rate": 0.00045395577733472087, + "loss": 3.636434316635132, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.0201494712382555, + "learning_rate": 0.0004537261325692383, + "loss": 3.679147720336914, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.02178201824426651, + "learning_rate": 0.0004534963663664332, + "loss": 3.671905040740967, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.024073559790849686, + "learning_rate": 0.00045326647891142394, + "loss": 3.6701040267944336, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.02135559171438217, + "learning_rate": 0.0004530364703894268, + "loss": 3.648519992828369, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.019264448434114456, + "learning_rate": 0.0004528063409857554, + "loss": 3.6985023021698, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.01922295056283474, + "learning_rate": 0.0004525760908858209, + "loss": 3.651756763458252, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.01978331431746483, + "learning_rate": 0.00045234572027513174, + "loss": 3.679884910583496, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.01968599483370781, + "learning_rate": 0.00045211522933929317, + "loss": 3.6571812629699707, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.020741639658808708, + "learning_rate": 0.0004518846182640077, + "loss": 3.6505298614501953, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.01842017099261284, + "learning_rate": 0.0004516538872350744, + "loss": 3.6904940605163574, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.018004486337304115, + "learning_rate": 0.0004514230364383893, + "loss": 3.7282047271728516, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.018821008503437042, + "learning_rate": 0.0004511920660599444, + "loss": 3.668164014816284, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.018929854035377502, + "learning_rate": 0.0004509609762858287, + "loss": 3.6672980785369873, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.01979168877005577, + "learning_rate": 0.0004507297673022269, + "loss": 3.6684999465942383, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.01774384267628193, + "learning_rate": 0.0004504984392954199, + "loss": 3.6577749252319336, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01876000687479973, + "learning_rate": 0.00045026699245178444, + "loss": 3.6629252433776855, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.018197249621152878, + "learning_rate": 0.0004500354269577932, + "loss": 3.688983678817749, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.019496265798807144, + "learning_rate": 0.00044980374300001414, + "loss": 3.6602022647857666, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01825535111129284, + "learning_rate": 0.00044957194076511095, + "loss": 3.6889238357543945, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.017113137990236282, + "learning_rate": 0.0004493400204398426, + "loss": 3.684070587158203, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.017993126064538956, + "learning_rate": 0.00044910798221106283, + "loss": 3.675426721572876, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.018354708328843117, + "learning_rate": 0.00044887582626572105, + "loss": 3.675588607788086, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.016362832859158516, + "learning_rate": 0.0004486435527908606, + "loss": 3.6709303855895996, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01772182621061802, + "learning_rate": 0.0004484111619736205, + "loss": 3.6884102821350098, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.019271764904260635, + "learning_rate": 0.0004481786540012336, + "loss": 3.6468968391418457, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.019040528684854507, + "learning_rate": 0.00044794602906102747, + "loss": 3.690387010574341, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.019022610038518906, + "learning_rate": 0.00044771328734042367, + "loss": 3.689873218536377, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.018533391878008842, + "learning_rate": 0.00044748042902693815, + "loss": 3.6877284049987793, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.020825615152716637, + "learning_rate": 0.0004472474543081805, + "loss": 3.662257671356201, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.021880341693758965, + "learning_rate": 0.00044701436337185444, + "loss": 3.650700092315674, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.019974704831838608, + "learning_rate": 0.00044678115640575704, + "loss": 3.684177875518799, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.01903127320110798, + "learning_rate": 0.0004465478335977788, + "loss": 3.658148765563965, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.020999105647206306, + "learning_rate": 0.00044631439513590393, + "loss": 3.68747615814209, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.02138398587703705, + "learning_rate": 0.00044608084120820933, + "loss": 3.656355381011963, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.02216442860662937, + "learning_rate": 0.0004458471720028655, + "loss": 3.6860342025756836, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.019359121099114418, + "learning_rate": 0.0004456133877081352, + "loss": 3.654944658279419, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.020984770730137825, + "learning_rate": 0.0004453794885123744, + "loss": 3.6885194778442383, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.02423972077667713, + "learning_rate": 0.00044514547460403135, + "loss": 3.672712802886963, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.024544453248381615, + "learning_rate": 0.00044491134617164694, + "loss": 3.6484768390655518, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.060352325439453, + "eval_runtime": 57.9556, + "eval_samples_per_second": 42.136, + "eval_steps_per_second": 1.329, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.024072550237178802, + "learning_rate": 0.000444677103403854, + "loss": 3.5796468257904053, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.028701433911919594, + "learning_rate": 0.0004444427464893779, + "loss": 3.604318618774414, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.028119578957557678, + "learning_rate": 0.00044420827561703573, + "loss": 3.608759641647339, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.025335192680358887, + "learning_rate": 0.0004439736909757365, + "loss": 3.5760319232940674, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.0273929201066494, + "learning_rate": 0.0004437389927544805, + "loss": 3.559441089630127, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.026567714288830757, + "learning_rate": 0.0004435041811423603, + "loss": 3.626889228820801, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.027587885037064552, + "learning_rate": 0.0004432692563285591, + "loss": 3.6226272583007812, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.02935059368610382, + "learning_rate": 0.0004430342185023517, + "loss": 3.601921558380127, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.02982175536453724, + "learning_rate": 0.0004427990678531038, + "loss": 3.59737491607666, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.028570258989930153, + "learning_rate": 0.0004425638045702719, + "loss": 3.56337833404541, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.029890350997447968, + "learning_rate": 0.0004423284288434035, + "loss": 3.6294307708740234, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.03206814453005791, + "learning_rate": 0.00044209294086213657, + "loss": 3.590505599975586, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.03169345110654831, + "learning_rate": 0.00044185734081619957, + "loss": 3.581584930419922, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.03313758969306946, + "learning_rate": 0.000441621628895411, + "loss": 3.6088438034057617, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.03317603841423988, + "learning_rate": 0.00044138580528967986, + "loss": 3.6340456008911133, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.034738484770059586, + "learning_rate": 0.0004411498701890051, + "loss": 3.5814096927642822, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.036428093910217285, + "learning_rate": 0.0004409138237834751, + "loss": 3.5939526557922363, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.03293406963348389, + "learning_rate": 0.00044067766626326835, + "loss": 3.577897787094116, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.031030720099806786, + "learning_rate": 0.00044044139781865255, + "loss": 3.6077990531921387, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.03441236540675163, + "learning_rate": 0.00044020501863998516, + "loss": 3.591923713684082, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.03500965237617493, + "learning_rate": 0.00043996852891771236, + "loss": 3.6171927452087402, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.034443050622940063, + "learning_rate": 0.00043973192884236983, + "loss": 3.619724750518799, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.03435051813721657, + "learning_rate": 0.0004394952186045819, + "loss": 3.6443169116973877, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.03230702877044678, + "learning_rate": 0.00043925839839506167, + "loss": 3.5912671089172363, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.03232988342642784, + "learning_rate": 0.0004390214684046111, + "loss": 3.618985414505005, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.029298555105924606, + "learning_rate": 0.00043878442882412015, + "loss": 3.6166677474975586, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.026723450049757957, + "learning_rate": 0.00043854727984456746, + "loss": 3.61362886428833, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.028280874714255333, + "learning_rate": 0.0004383100216570197, + "loss": 3.6410951614379883, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.026700228452682495, + "learning_rate": 0.0004380726544526316, + "loss": 3.5804049968719482, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.023092713207006454, + "learning_rate": 0.00043783517842264534, + "loss": 3.595339298248291, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.024123897776007652, + "learning_rate": 0.0004375975937583914, + "loss": 3.647569417953491, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.02505572699010372, + "learning_rate": 0.0004373599006512871, + "loss": 3.6276652812957764, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.024523509666323662, + "learning_rate": 0.00043712209929283786, + "loss": 3.59525728225708, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.024351581931114197, + "learning_rate": 0.00043688418987463567, + "loss": 3.6081674098968506, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.02530614472925663, + "learning_rate": 0.00043664617258835995, + "loss": 3.6632542610168457, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.023438062518835068, + "learning_rate": 0.0004364080476257769, + "loss": 3.6345431804656982, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.021853657439351082, + "learning_rate": 0.0004361698151787396, + "loss": 3.598644256591797, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.022487422451376915, + "learning_rate": 0.0004359314754391874, + "loss": 3.5966577529907227, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.021353935822844505, + "learning_rate": 0.0004356930285991466, + "loss": 3.6328375339508057, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.021180452778935432, + "learning_rate": 0.00043545447485072926, + "loss": 3.6285812854766846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.020491832867264748, + "learning_rate": 0.000435215814386134, + "loss": 3.592944622039795, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.021603770554065704, + "learning_rate": 0.0004349770473976453, + "loss": 3.613507032394409, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.02146158367395401, + "learning_rate": 0.0004347381740776332, + "loss": 3.6409506797790527, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.02102433145046234, + "learning_rate": 0.0004344991946185538, + "loss": 3.6179330348968506, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.018684551119804382, + "learning_rate": 0.00043426010921294837, + "loss": 3.6385130882263184, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.018416227772831917, + "learning_rate": 0.0004340209180534438, + "loss": 3.642151355743408, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.01915144734084606, + "learning_rate": 0.000433781621332752, + "loss": 3.6089656352996826, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.018299665302038193, + "learning_rate": 0.0004335422192436703, + "loss": 3.6063547134399414, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.018962757661938667, + "learning_rate": 0.00043330271197908023, + "loss": 3.625917434692383, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.017437821254134178, + "learning_rate": 0.00043306309973194874, + "loss": 3.600015640258789, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.017913147807121277, + "learning_rate": 0.000432823382695327, + "loss": 3.6428744792938232, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.018343618139624596, + "learning_rate": 0.0004325835610623508, + "loss": 3.635901927947998, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.020372850820422173, + "learning_rate": 0.00043234363502623994, + "loss": 3.620181083679199, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.02037600241601467, + "learning_rate": 0.0004321036047802985, + "loss": 3.61129093170166, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.021827151998877525, + "learning_rate": 0.00043186347051791466, + "loss": 3.5999555587768555, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.021806657314300537, + "learning_rate": 0.0004316232324325602, + "loss": 3.642509937286377, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02050699107348919, + "learning_rate": 0.0004313828907177906, + "loss": 3.597583293914795, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.020436758175492287, + "learning_rate": 0.0004311424455672448, + "loss": 3.6079912185668945, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.019341666251420975, + "learning_rate": 0.0004309018971746451, + "loss": 3.6196017265319824, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.022348817437887192, + "learning_rate": 0.000430661245733797, + "loss": 3.619025468826294, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.019361121580004692, + "learning_rate": 0.00043042049143858914, + "loss": 3.657182455062866, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.019258935004472733, + "learning_rate": 0.0004301796344829928, + "loss": 3.6408963203430176, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.020622272044420242, + "learning_rate": 0.00042993867506106205, + "loss": 3.6247220039367676, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.019510755315423012, + "learning_rate": 0.0004296976133669336, + "loss": 3.6264584064483643, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.02078665979206562, + "learning_rate": 0.0004294564495948266, + "loss": 3.6364388465881348, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01791800744831562, + "learning_rate": 0.0004292151839390421, + "loss": 3.6029374599456787, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.01922667771577835, + "learning_rate": 0.0004289738165939636, + "loss": 3.582937717437744, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.018473893404006958, + "learning_rate": 0.00042873234775405654, + "loss": 3.6141819953918457, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01846298575401306, + "learning_rate": 0.00042849077761386784, + "loss": 3.638786554336548, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.01876973733305931, + "learning_rate": 0.0004282491063680263, + "loss": 3.617130756378174, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.017995499074459076, + "learning_rate": 0.00042800733421124207, + "loss": 3.645305633544922, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.02105429396033287, + "learning_rate": 0.00042776546133830646, + "loss": 3.6444876194000244, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.049062252044678, + "eval_runtime": 58.0113, + "eval_samples_per_second": 42.095, + "eval_steps_per_second": 1.327, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.02128119394183159, + "learning_rate": 0.00042752348794409227, + "loss": 3.524954080581665, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.028249500319361687, + "learning_rate": 0.00042728141422355287, + "loss": 3.5311479568481445, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.03145391866564751, + "learning_rate": 0.0004270392403717229, + "loss": 3.559330940246582, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.02994118444621563, + "learning_rate": 0.00042679696658371743, + "loss": 3.5257620811462402, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.027448246255517006, + "learning_rate": 0.00042655459305473195, + "loss": 3.5242764949798584, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.0296134352684021, + "learning_rate": 0.0004263121199800425, + "loss": 3.5317091941833496, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.029576804488897324, + "learning_rate": 0.0004260695475550054, + "loss": 3.5340428352355957, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.02941269613802433, + "learning_rate": 0.0004258268759750566, + "loss": 3.5372023582458496, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.03126693144440651, + "learning_rate": 0.0004255841054357124, + "loss": 3.5437135696411133, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.03335367515683174, + "learning_rate": 0.0004253412361325686, + "loss": 3.5598273277282715, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.03028946928679943, + "learning_rate": 0.0004250982682613006, + "loss": 3.5190839767456055, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.028629133477807045, + "learning_rate": 0.00042485520201766315, + "loss": 3.5138821601867676, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.027070026844739914, + "learning_rate": 0.0004246120375974905, + "loss": 3.530958414077759, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.028837760910391808, + "learning_rate": 0.0004243687751966957, + "loss": 3.5233962535858154, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.029884297400712967, + "learning_rate": 0.0004241254150112708, + "loss": 3.5639758110046387, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.024223247542977333, + "learning_rate": 0.000423881957237287, + "loss": 3.5555152893066406, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.026992836967110634, + "learning_rate": 0.0004236384020708935, + "loss": 3.540196418762207, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.02818330004811287, + "learning_rate": 0.0004233947497083185, + "loss": 3.5546071529388428, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.026401042938232422, + "learning_rate": 0.0004231510003458682, + "loss": 3.5556247234344482, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.026020489633083344, + "learning_rate": 0.0004229071541799272, + "loss": 3.568004608154297, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.02803782932460308, + "learning_rate": 0.0004226632114069578, + "loss": 3.5828235149383545, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.02737978659570217, + "learning_rate": 0.0004224191722235004, + "loss": 3.5642499923706055, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.02406371757388115, + "learning_rate": 0.00042217503682617286, + "loss": 3.5530993938446045, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.02230692468583584, + "learning_rate": 0.0004219308054116706, + "loss": 3.5573501586914062, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.021352406591176987, + "learning_rate": 0.0004216864781767667, + "loss": 3.5526974201202393, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.02227999083697796, + "learning_rate": 0.00042144205531831103, + "loss": 3.5866928100585938, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.024190247058868408, + "learning_rate": 0.0004211975370332308, + "loss": 3.517117977142334, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.023196619004011154, + "learning_rate": 0.00042095292351852976, + "loss": 3.4826502799987793, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.02388714998960495, + "learning_rate": 0.0004207082149712888, + "loss": 3.5588626861572266, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.024692339822649956, + "learning_rate": 0.0004204634115886651, + "loss": 3.54828143119812, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.023078206926584244, + "learning_rate": 0.0004202185135678924, + "loss": 3.5679030418395996, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.020665735006332397, + "learning_rate": 0.00041997352110628033, + "loss": 3.5500283241271973, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.02022322081029415, + "learning_rate": 0.00041972843440121527, + "loss": 3.592104911804199, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.021137656643986702, + "learning_rate": 0.00041948325365015894, + "loss": 3.583833694458008, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.021026931703090668, + "learning_rate": 0.0004192379790506491, + "loss": 3.556403875350952, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.020423648878932, + "learning_rate": 0.00041899261080029904, + "loss": 3.5839173793792725, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.020935002714395523, + "learning_rate": 0.00041874714909679754, + "loss": 3.5794217586517334, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.020799759775400162, + "learning_rate": 0.00041850159413790863, + "loss": 3.6033859252929688, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.019480884075164795, + "learning_rate": 0.0004182559461214715, + "loss": 3.580097198486328, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.019300777465105057, + "learning_rate": 0.0004180102052454004, + "loss": 3.550474166870117, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.020265772938728333, + "learning_rate": 0.00041776437170768427, + "loss": 3.5711655616760254, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.019972780719399452, + "learning_rate": 0.0004175184457063868, + "loss": 3.5636885166168213, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.019635623320937157, + "learning_rate": 0.000417272427439646, + "loss": 3.5851311683654785, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.021451901644468307, + "learning_rate": 0.00041702631710567443, + "loss": 3.5956759452819824, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.019456414505839348, + "learning_rate": 0.00041678011490275875, + "loss": 3.595731258392334, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.019757689908146858, + "learning_rate": 0.00041653382102925957, + "loss": 3.569652557373047, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.020571604371070862, + "learning_rate": 0.00041628743568361147, + "loss": 3.6007747650146484, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.02153848111629486, + "learning_rate": 0.00041604095906432265, + "loss": 3.5915260314941406, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.020875850692391396, + "learning_rate": 0.00041579439136997493, + "loss": 3.5582690238952637, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.02089419960975647, + "learning_rate": 0.00041554773279922337, + "loss": 3.5756618976593018, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01961098238825798, + "learning_rate": 0.0004153009835507963, + "loss": 3.5918776988983154, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.021075716242194176, + "learning_rate": 0.0004150541438234952, + "loss": 3.5342202186584473, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.020745476707816124, + "learning_rate": 0.00041480721381619435, + "loss": 3.5976028442382812, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.02001352794468403, + "learning_rate": 0.00041456019372784086, + "loss": 3.5686182975769043, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.021684160456061363, + "learning_rate": 0.0004143130837574543, + "loss": 3.553445339202881, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.022551920264959335, + "learning_rate": 0.0004140658841041267, + "loss": 3.5712709426879883, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.022758224979043007, + "learning_rate": 0.0004138185949670224, + "loss": 3.6101667881011963, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.020714636892080307, + "learning_rate": 0.00041357121654537766, + "loss": 3.5740506649017334, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.022694187238812447, + "learning_rate": 0.00041332374903850104, + "loss": 3.5663979053497314, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.021664340049028397, + "learning_rate": 0.00041307619264577234, + "loss": 3.5830888748168945, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.019072270020842552, + "learning_rate": 0.0004128285475666436, + "loss": 3.5861973762512207, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.020153764635324478, + "learning_rate": 0.0004125808140006378, + "loss": 3.566444158554077, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019666558131575584, + "learning_rate": 0.0004123329921473494, + "loss": 3.5789592266082764, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.02101624570786953, + "learning_rate": 0.0004120850822064439, + "loss": 3.551722526550293, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.01951020397245884, + "learning_rate": 0.00041183708437765815, + "loss": 3.6069014072418213, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.01985476352274418, + "learning_rate": 0.00041158899886079925, + "loss": 3.5944924354553223, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02195335365831852, + "learning_rate": 0.00041134082585574534, + "loss": 3.5828046798706055, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02105172537267208, + "learning_rate": 0.00041109256556244497, + "loss": 3.5830953121185303, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.01888827420771122, + "learning_rate": 0.00041084421818091685, + "loss": 3.5656847953796387, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.017867228016257286, + "learning_rate": 0.0004105957839112501, + "loss": 3.5792243480682373, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.018577905371785164, + "learning_rate": 0.0004103472629536036, + "loss": 3.56992506980896, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.020045632496476173, + "learning_rate": 0.0004100986555082063, + "loss": 3.590876579284668, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.044339656829834, + "eval_runtime": 57.1435, + "eval_samples_per_second": 42.735, + "eval_steps_per_second": 1.347, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.024135645478963852, + "learning_rate": 0.0004098499617753567, + "loss": 3.4856467247009277, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.031053155660629272, + "learning_rate": 0.00040960118195542273, + "loss": 3.455773115158081, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.035190410912036896, + "learning_rate": 0.00040935231624884204, + "loss": 3.506619930267334, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.03611094877123833, + "learning_rate": 0.0004091033648561211, + "loss": 3.5227696895599365, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.034121073782444, + "learning_rate": 0.00040885432797783543, + "loss": 3.4970145225524902, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.03629382699728012, + "learning_rate": 0.0004086052058146297, + "loss": 3.506425619125366, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.03709045797586441, + "learning_rate": 0.00040835599856721725, + "loss": 3.499274969100952, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.034826792776584625, + "learning_rate": 0.00040810670643637954, + "loss": 3.5029163360595703, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.03344690054655075, + "learning_rate": 0.000407857329622967, + "loss": 3.4816911220550537, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.03136949986219406, + "learning_rate": 0.00040760786832789786, + "loss": 3.5154659748077393, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.030533097684383392, + "learning_rate": 0.0004073583227521588, + "loss": 3.509666919708252, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.03298318013548851, + "learning_rate": 0.0004071086930968039, + "loss": 3.5096616744995117, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.031186232343316078, + "learning_rate": 0.00040685897956295545, + "loss": 3.4752590656280518, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.028008371591567993, + "learning_rate": 0.0004066091823518031, + "loss": 3.498826503753662, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.028206519782543182, + "learning_rate": 0.00040635930166460385, + "loss": 3.5281882286071777, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.024417119100689888, + "learning_rate": 0.00040610933770268226, + "loss": 3.506791830062866, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.025321070104837418, + "learning_rate": 0.00040585929066742964, + "loss": 3.494256019592285, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.023430505767464638, + "learning_rate": 0.00040560916076030435, + "loss": 3.533010482788086, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.023371374234557152, + "learning_rate": 0.00040535894818283154, + "loss": 3.5197086334228516, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.02414027415215969, + "learning_rate": 0.0004051086531366031, + "loss": 3.4707889556884766, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.026652060449123383, + "learning_rate": 0.000404858275823277, + "loss": 3.498159885406494, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.024486590176820755, + "learning_rate": 0.000404607816444578, + "loss": 3.50034236907959, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.023153992369771004, + "learning_rate": 0.00040435727520229644, + "loss": 3.499504804611206, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.02222427912056446, + "learning_rate": 0.0004041066522982891, + "loss": 3.4953579902648926, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.023033641278743744, + "learning_rate": 0.000403855947934478, + "loss": 3.5181496143341064, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.02141181193292141, + "learning_rate": 0.00040360516231285155, + "loss": 3.552860736846924, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.023448696359992027, + "learning_rate": 0.00040335429563546286, + "loss": 3.5153324604034424, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.022410666570067406, + "learning_rate": 0.0004031033481044308, + "loss": 3.4963512420654297, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.020326802507042885, + "learning_rate": 0.00040285231992193924, + "loss": 3.528319835662842, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021117394790053368, + "learning_rate": 0.0004026012112902372, + "loss": 3.5192837715148926, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.021317072212696075, + "learning_rate": 0.0004023500224116381, + "loss": 3.5081021785736084, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.021116144955158234, + "learning_rate": 0.00040209875348852037, + "loss": 3.5167174339294434, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019834019243717194, + "learning_rate": 0.00040184740472332705, + "loss": 3.511960983276367, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.02089378982782364, + "learning_rate": 0.000401595976318565, + "loss": 3.493330955505371, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.021129710599780083, + "learning_rate": 0.00040134446847680584, + "loss": 3.544156551361084, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.021163299679756165, + "learning_rate": 0.0004010928814006846, + "loss": 3.5261003971099854, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.02105306088924408, + "learning_rate": 0.0004008412152929007, + "loss": 3.543973445892334, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.02150859497487545, + "learning_rate": 0.0004005894703562168, + "loss": 3.487217426300049, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.021666452288627625, + "learning_rate": 0.0004003376467934593, + "loss": 3.4933671951293945, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.02180723287165165, + "learning_rate": 0.0004000857448075178, + "loss": 3.55198073387146, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.02041625790297985, + "learning_rate": 0.0003998337646013455, + "loss": 3.529571533203125, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.022977063432335854, + "learning_rate": 0.00039958170637795806, + "loss": 3.526432752609253, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.022725161164999008, + "learning_rate": 0.0003993295703404341, + "loss": 3.5258829593658447, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.021337240934371948, + "learning_rate": 0.00039907735669191544, + "loss": 3.4897351264953613, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.020783131942152977, + "learning_rate": 0.00039882506563560573, + "loss": 3.5497589111328125, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.022259317338466644, + "learning_rate": 0.0003985726973747715, + "loss": 3.5225064754486084, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.020283272489905357, + "learning_rate": 0.00039832025211274113, + "loss": 3.5453078746795654, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.020656265318393707, + "learning_rate": 0.00039806773005290544, + "loss": 3.533703327178955, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.02104976586997509, + "learning_rate": 0.0003978151313987168, + "loss": 3.4885146617889404, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.021411150693893433, + "learning_rate": 0.00039756245635368933, + "loss": 3.533268690109253, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.021906673908233643, + "learning_rate": 0.00039730970512139876, + "loss": 3.5483224391937256, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.02415728010237217, + "learning_rate": 0.0003970568779054822, + "loss": 3.5084290504455566, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.023435061797499657, + "learning_rate": 0.000396803974909638, + "loss": 3.5498428344726562, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.02183385379612446, + "learning_rate": 0.0003965509963376255, + "loss": 3.484086751937866, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.022856134921312332, + "learning_rate": 0.000396297942393265, + "loss": 3.5339231491088867, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.021679896861314774, + "learning_rate": 0.0003960448132804375, + "loss": 3.547191619873047, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.021262124180793762, + "learning_rate": 0.0003957916092030845, + "loss": 3.5263822078704834, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.02265402302145958, + "learning_rate": 0.00039553833036520803, + "loss": 3.5361135005950928, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.020080721005797386, + "learning_rate": 0.0003952849769708702, + "loss": 3.5247459411621094, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.02187187783420086, + "learning_rate": 0.0003950315492241932, + "loss": 3.54129695892334, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.02417500875890255, + "learning_rate": 0.0003947780473293593, + "loss": 3.598796844482422, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.02416270412504673, + "learning_rate": 0.00039452447149061054, + "loss": 3.5207936763763428, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.021958643570542336, + "learning_rate": 0.0003942708219122481, + "loss": 3.5395452976226807, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.021813420578837395, + "learning_rate": 0.000394017098798633, + "loss": 3.5469255447387695, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020607495680451393, + "learning_rate": 0.0003937633023541854, + "loss": 3.554851531982422, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.02091527171432972, + "learning_rate": 0.0003935094327833845, + "loss": 3.492576837539673, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.021089257672429085, + "learning_rate": 0.0003932554902907683, + "loss": 3.5154709815979004, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.019659532234072685, + "learning_rate": 0.0003930014750809338, + "loss": 3.5444283485412598, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01896163448691368, + "learning_rate": 0.0003927473873585365, + "loss": 3.589625358581543, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020288823172450066, + "learning_rate": 0.0003924932273282903, + "loss": 3.522411823272705, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.020222362130880356, + "learning_rate": 0.00039223899519496723, + "loss": 3.5244359970092773, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.022349944338202477, + "learning_rate": 0.00039198469116339757, + "loss": 3.5362396240234375, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.047028064727783, + "eval_runtime": 62.8948, + "eval_samples_per_second": 38.827, + "eval_steps_per_second": 1.224, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.02202913723886013, + "learning_rate": 0.0003917303154384694, + "loss": 3.4249138832092285, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.03153969347476959, + "learning_rate": 0.00039147586822512885, + "loss": 3.4436073303222656, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.03747870773077011, + "learning_rate": 0.0003912213497283793, + "loss": 3.4879579544067383, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.036144088953733444, + "learning_rate": 0.0003909667601532819, + "loss": 3.419151782989502, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.03983336314558983, + "learning_rate": 0.00039071209970495445, + "loss": 3.4331436157226562, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.045912645757198334, + "learning_rate": 0.00039045736858857273, + "loss": 3.4304769039154053, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.04005753993988037, + "learning_rate": 0.0003902025670093687, + "loss": 3.478600025177002, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.03488151356577873, + "learning_rate": 0.00038994769517263156, + "loss": 3.4650232791900635, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.03603382781147957, + "learning_rate": 0.0003896927532837069, + "loss": 3.449807643890381, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.03817906603217125, + "learning_rate": 0.0003894377415479967, + "loss": 3.4556589126586914, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.03523671627044678, + "learning_rate": 0.00038918266017095936, + "loss": 3.477426767349243, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.03284068778157234, + "learning_rate": 0.0003889275093581096, + "loss": 3.478672981262207, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.02992352657020092, + "learning_rate": 0.00038867228931501737, + "loss": 3.4810047149658203, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.02858242578804493, + "learning_rate": 0.0003884170002473091, + "loss": 3.467811107635498, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.027912171557545662, + "learning_rate": 0.0003881616423606666, + "loss": 3.4326171875, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.028351375833153725, + "learning_rate": 0.00038790621586082697, + "loss": 3.4533729553222656, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.027947887778282166, + "learning_rate": 0.0003876507209535829, + "loss": 3.420295476913452, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.028510035946965218, + "learning_rate": 0.00038739515784478186, + "loss": 3.438023328781128, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.027707934379577637, + "learning_rate": 0.00038713952674032654, + "loss": 3.4844181537628174, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.0262764822691679, + "learning_rate": 0.0003868838278461743, + "loss": 3.4363749027252197, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.02713681012392044, + "learning_rate": 0.0003866280613683371, + "loss": 3.4501900672912598, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.0299485195428133, + "learning_rate": 0.00038637222751288126, + "loss": 3.461942672729492, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.027047986164689064, + "learning_rate": 0.0003861163264859277, + "loss": 3.454293727874756, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.02436736971139908, + "learning_rate": 0.0003858603584936511, + "loss": 3.4663405418395996, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.02601189725100994, + "learning_rate": 0.0003856043237422803, + "loss": 3.502030849456787, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.02463468722999096, + "learning_rate": 0.0003853482224380978, + "loss": 3.462491035461426, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.021793190389871597, + "learning_rate": 0.00038509205478744, + "loss": 3.4476377964019775, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.023809107020497322, + "learning_rate": 0.0003848358209966963, + "loss": 3.4701178073883057, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.02557182125747204, + "learning_rate": 0.0003845795212723099, + "loss": 3.4932539463043213, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.023424552753567696, + "learning_rate": 0.00038432315582077664, + "loss": 3.4779105186462402, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.02195148915052414, + "learning_rate": 0.0003840667248486456, + "loss": 3.470506191253662, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.022897250950336456, + "learning_rate": 0.0003838102285625186, + "loss": 3.48982572555542, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.023033805191516876, + "learning_rate": 0.00038355366716905006, + "loss": 3.4845101833343506, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.022774485871195793, + "learning_rate": 0.0003832970408749467, + "loss": 3.5060324668884277, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.02193611115217209, + "learning_rate": 0.00038304034988696816, + "loss": 3.511518955230713, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.02130570448935032, + "learning_rate": 0.00038278359441192516, + "loss": 3.468502998352051, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.020402755588293076, + "learning_rate": 0.00038252677465668136, + "loss": 3.4786908626556396, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.022664198651909828, + "learning_rate": 0.00038226989082815156, + "loss": 3.479928970336914, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.02454056777060032, + "learning_rate": 0.00038201294313330245, + "loss": 3.479947566986084, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.022096732631325722, + "learning_rate": 0.00038175593177915227, + "loss": 3.49478816986084, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.021499406546354294, + "learning_rate": 0.0003814988569727704, + "loss": 3.494933605194092, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.022325703874230385, + "learning_rate": 0.0003812417189212773, + "loss": 3.5041463375091553, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.020674455910921097, + "learning_rate": 0.00038098451783184447, + "loss": 3.472404718399048, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.02131054364144802, + "learning_rate": 0.00038072725391169416, + "loss": 3.497795343399048, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.02126218006014824, + "learning_rate": 0.0003804699273680994, + "loss": 3.483281373977661, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.0217830128967762, + "learning_rate": 0.0003802125384083834, + "loss": 3.4704298973083496, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.02156917378306389, + "learning_rate": 0.0003799550872399197, + "loss": 3.4994399547576904, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.02047055773437023, + "learning_rate": 0.0003796975740701323, + "loss": 3.4316084384918213, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.019766805693507195, + "learning_rate": 0.0003794399991064948, + "loss": 3.452223777770996, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.020028885453939438, + "learning_rate": 0.00037918236255653074, + "loss": 3.485278844833374, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.020312055945396423, + "learning_rate": 0.00037892466462781306, + "loss": 3.4789156913757324, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.020725280046463013, + "learning_rate": 0.00037866690552796446, + "loss": 3.495528221130371, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.02088077738881111, + "learning_rate": 0.00037840908546465696, + "loss": 3.4972569942474365, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.021010037511587143, + "learning_rate": 0.0003781512046456111, + "loss": 3.473562240600586, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.019641095772385597, + "learning_rate": 0.0003778932632785972, + "loss": 3.4782915115356445, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.02063329517841339, + "learning_rate": 0.00037763526157143376, + "loss": 3.4681782722473145, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.0213928259909153, + "learning_rate": 0.00037737719973198825, + "loss": 3.472263813018799, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.02135022170841694, + "learning_rate": 0.0003771190779681762, + "loss": 3.502300262451172, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.02053583413362503, + "learning_rate": 0.0003768608964879619, + "loss": 3.489243745803833, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.021575238555669785, + "learning_rate": 0.00037660265549935724, + "loss": 3.491074562072754, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.022335771471261978, + "learning_rate": 0.00037634435521042257, + "loss": 3.5298843383789062, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.021198464557528496, + "learning_rate": 0.0003760859958292656, + "loss": 3.5156898498535156, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.020986037328839302, + "learning_rate": 0.000375827577564042, + "loss": 3.4574475288391113, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.022639811038970947, + "learning_rate": 0.0003755691006229545, + "loss": 3.4952611923217773, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.019711066037416458, + "learning_rate": 0.0003753105652142534, + "loss": 3.531296730041504, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.02058061584830284, + "learning_rate": 0.000375051971546236, + "loss": 3.5340657234191895, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.019849255681037903, + "learning_rate": 0.0003747933198272465, + "loss": 3.5028295516967773, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01954430714249611, + "learning_rate": 0.00037453461026567604, + "loss": 3.51088809967041, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.019234873354434967, + "learning_rate": 0.00037427584306996196, + "loss": 3.472560167312622, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.021085912361741066, + "learning_rate": 0.0003740170184485888, + "loss": 3.468876838684082, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.020300406962633133, + "learning_rate": 0.0003737581366100864, + "loss": 3.4903886318206787, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02380351349711418, + "learning_rate": 0.0003734991977630315, + "loss": 3.5069704055786133, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.048111915588379, + "eval_runtime": 58.312, + "eval_samples_per_second": 41.878, + "eval_steps_per_second": 1.32, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02379327267408371, + "learning_rate": 0.0003732402021160463, + "loss": 3.396376132965088, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.03204222396016121, + "learning_rate": 0.00037298114987779885, + "loss": 3.4102582931518555, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.034993767738342285, + "learning_rate": 0.000372722041257003, + "loss": 3.385697364807129, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.03387003391981125, + "learning_rate": 0.00037246287646241783, + "loss": 3.4120078086853027, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.03282478079199791, + "learning_rate": 0.0003722036557028478, + "loss": 3.404122829437256, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.03053346462547779, + "learning_rate": 0.0003719443791871422, + "loss": 3.368051767349243, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.029722390696406364, + "learning_rate": 0.0003716850471241958, + "loss": 3.4066247940063477, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.0287226140499115, + "learning_rate": 0.0003714256597229474, + "loss": 3.4212646484375, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.028156550601124763, + "learning_rate": 0.0003711662171923809, + "loss": 3.4192593097686768, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.03021939843893051, + "learning_rate": 0.0003709067197415244, + "loss": 3.3973188400268555, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.03314025700092316, + "learning_rate": 0.00037064716757945036, + "loss": 3.422978401184082, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.03156686946749687, + "learning_rate": 0.0003703875609152753, + "loss": 3.4052343368530273, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.03357478231191635, + "learning_rate": 0.0003701278999581595, + "loss": 3.383730411529541, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.03261081501841545, + "learning_rate": 0.0003698681849173073, + "loss": 3.421250820159912, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.03252755105495453, + "learning_rate": 0.0003696084160019662, + "loss": 3.39884352684021, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.03177274018526077, + "learning_rate": 0.0003693485934214274, + "loss": 3.4175422191619873, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.028845487162470818, + "learning_rate": 0.0003690887173850253, + "loss": 3.3908143043518066, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.026045162230730057, + "learning_rate": 0.0003688287881021374, + "loss": 3.4417710304260254, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.027937524020671844, + "learning_rate": 0.00036856880578218376, + "loss": 3.410548210144043, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.025780271738767624, + "learning_rate": 0.00036830877063462786, + "loss": 3.4069018363952637, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.026093080639839172, + "learning_rate": 0.0003680486828689749, + "loss": 3.4199600219726562, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.025765180587768555, + "learning_rate": 0.00036778854269477315, + "loss": 3.442903995513916, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.022773101925849915, + "learning_rate": 0.0003675283503216127, + "loss": 3.394232749938965, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.025135623291134834, + "learning_rate": 0.000367268105959126, + "loss": 3.4466171264648438, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.02491890639066696, + "learning_rate": 0.00036700780981698705, + "loss": 3.418689727783203, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.02521841786801815, + "learning_rate": 0.0003667474621049119, + "loss": 3.4213061332702637, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.025186199694871902, + "learning_rate": 0.00036648706303265795, + "loss": 3.437631130218506, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.022097431123256683, + "learning_rate": 0.000366226612810024, + "loss": 3.4014902114868164, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.024009618908166885, + "learning_rate": 0.00036596611164685025, + "loss": 3.4292187690734863, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.022758696228265762, + "learning_rate": 0.0003657055597530175, + "loss": 3.4105381965637207, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.02447321265935898, + "learning_rate": 0.00036544495733844797, + "loss": 3.4196958541870117, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.025843804702162743, + "learning_rate": 0.000365184304613104, + "loss": 3.4343783855438232, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.022433951497077942, + "learning_rate": 0.0003649236017869892, + "loss": 3.440580368041992, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.02250702679157257, + "learning_rate": 0.00036466284907014685, + "loss": 3.4255261421203613, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.022953296080231667, + "learning_rate": 0.0003644020466726608, + "loss": 3.442258358001709, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.021977661177515984, + "learning_rate": 0.00036414119480465473, + "loss": 3.4058687686920166, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.021751178428530693, + "learning_rate": 0.0003638802936762925, + "loss": 3.4211649894714355, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.022764205932617188, + "learning_rate": 0.0003636193434977771, + "loss": 3.424694061279297, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.024827787652611732, + "learning_rate": 0.0003633583444793516, + "loss": 3.458076238632202, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.021888602524995804, + "learning_rate": 0.00036309729683129814, + "loss": 3.4509153366088867, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.021961702033877373, + "learning_rate": 0.00036283620076393814, + "loss": 3.4147253036499023, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.02235778421163559, + "learning_rate": 0.000362575056487632, + "loss": 3.4392640590667725, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.021714260801672935, + "learning_rate": 0.00036231386421277877, + "loss": 3.444026470184326, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.020996905863285065, + "learning_rate": 0.0003620526241498167, + "loss": 3.4168434143066406, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.020528919994831085, + "learning_rate": 0.0003617913365092218, + "loss": 3.4831130504608154, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.021056141704320908, + "learning_rate": 0.0003615300015015091, + "loss": 3.430905342102051, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.021655110642313957, + "learning_rate": 0.0003612686193372312, + "loss": 3.4588241577148438, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.02344612590968609, + "learning_rate": 0.00036100719022697924, + "loss": 3.4202680587768555, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.02264723740518093, + "learning_rate": 0.0003607457143813819, + "loss": 3.4669015407562256, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.02377854660153389, + "learning_rate": 0.00036048419201110553, + "loss": 3.5059216022491455, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.02299923449754715, + "learning_rate": 0.00036022262332685383, + "loss": 3.4796524047851562, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.021110547706484795, + "learning_rate": 0.0003599610085393681, + "loss": 3.4342408180236816, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.022060122340917587, + "learning_rate": 0.0003596993478594267, + "loss": 3.426923990249634, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.021077830344438553, + "learning_rate": 0.0003594376414978447, + "loss": 3.4636270999908447, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.02209082990884781, + "learning_rate": 0.0003591758896654745, + "loss": 3.441979169845581, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.022408071905374527, + "learning_rate": 0.0003589140925732045, + "loss": 3.490835666656494, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.022462666034698486, + "learning_rate": 0.0003586522504319602, + "loss": 3.452770233154297, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.021295800805091858, + "learning_rate": 0.0003583903634527029, + "loss": 3.4377408027648926, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.020457757636904716, + "learning_rate": 0.0003581284318464302, + "loss": 3.4191250801086426, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.021163633093237877, + "learning_rate": 0.00035786645582417564, + "loss": 3.4797091484069824, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.021495461463928223, + "learning_rate": 0.00035760443559700863, + "loss": 3.4763145446777344, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.021050071343779564, + "learning_rate": 0.00035734237137603417, + "loss": 3.4687209129333496, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.02150929905474186, + "learning_rate": 0.0003570802633723927, + "loss": 3.4819254875183105, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.0212967898696661, + "learning_rate": 0.0003568181117972597, + "loss": 3.490407705307007, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.02123691327869892, + "learning_rate": 0.00035655591686184626, + "loss": 3.4418535232543945, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.02193271741271019, + "learning_rate": 0.0003562936787773979, + "loss": 3.446892738342285, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.02065582573413849, + "learning_rate": 0.000356031397755195, + "loss": 3.4308691024780273, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.019487040117383003, + "learning_rate": 0.00035576907400655303, + "loss": 3.487239122390747, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.021190891042351723, + "learning_rate": 0.00035550670774282106, + "loss": 3.4635558128356934, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.0209872554987669, + "learning_rate": 0.0003552442991753831, + "loss": 3.443793773651123, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.02035645768046379, + "learning_rate": 0.00035498184851565697, + "loss": 3.473198652267456, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.024002201855182648, + "learning_rate": 0.0003547193559750944, + "loss": 3.4244751930236816, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.057085990905762, + "eval_runtime": 62.4735, + "eval_samples_per_second": 39.089, + "eval_steps_per_second": 1.233, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.02450242079794407, + "learning_rate": 0.00035445682176518087, + "loss": 3.361237049102783, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.030813416466116905, + "learning_rate": 0.00035419424609743546, + "loss": 3.365708351135254, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.03279578685760498, + "learning_rate": 0.00035393162918341065, + "loss": 3.325385570526123, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.03293970227241516, + "learning_rate": 0.00035366897123469213, + "loss": 3.356811285018921, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.033190157264471054, + "learning_rate": 0.0003534062724628986, + "loss": 3.406217098236084, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.033370476216077805, + "learning_rate": 0.00035314353307968195, + "loss": 3.3579213619232178, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.03351330757141113, + "learning_rate": 0.00035288075329672634, + "loss": 3.3755106925964355, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.034596871584653854, + "learning_rate": 0.0003526179333257487, + "loss": 3.3576536178588867, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.03177076578140259, + "learning_rate": 0.0003523550733784984, + "loss": 3.3123257160186768, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.0289607010781765, + "learning_rate": 0.00035209217366675706, + "loss": 3.370959997177124, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.03265267238020897, + "learning_rate": 0.000351829234402338, + "loss": 3.332608938217163, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.03397805988788605, + "learning_rate": 0.0003515662557970867, + "loss": 3.364213228225708, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.03100395016372204, + "learning_rate": 0.0003513032380628804, + "loss": 3.388005256652832, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.03224753588438034, + "learning_rate": 0.0003510401814116277, + "loss": 3.3267595767974854, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.03034576214849949, + "learning_rate": 0.0003507770860552684, + "loss": 3.3954129219055176, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.026316817849874496, + "learning_rate": 0.00035051395220577397, + "loss": 3.3543832302093506, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.027714375406503677, + "learning_rate": 0.0003502507800751464, + "loss": 3.4190099239349365, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.028731998056173325, + "learning_rate": 0.0003499875698754187, + "loss": 3.342031955718994, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02845810353755951, + "learning_rate": 0.00034972432181865467, + "loss": 3.356130599975586, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.02842753194272518, + "learning_rate": 0.00034946103611694854, + "loss": 3.3709869384765625, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.02942183054983616, + "learning_rate": 0.00034919771298242477, + "loss": 3.373588800430298, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.02958574891090393, + "learning_rate": 0.00034893435262723814, + "loss": 3.377514600753784, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.029524607583880424, + "learning_rate": 0.00034867095526357325, + "loss": 3.398663282394409, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.028092598542571068, + "learning_rate": 0.0003484075211036446, + "loss": 3.3742518424987793, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.026652829721570015, + "learning_rate": 0.0003481440503596964, + "loss": 3.387667179107666, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.024675756692886353, + "learning_rate": 0.0003478805432440021, + "loss": 3.3737070560455322, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.02491692826151848, + "learning_rate": 0.0003476169999688648, + "loss": 3.359217643737793, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.024601681157946587, + "learning_rate": 0.00034735342074661654, + "loss": 3.3902688026428223, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.02313913404941559, + "learning_rate": 0.0003470898057896183, + "loss": 3.3700339794158936, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.02418898604810238, + "learning_rate": 0.0003468261553102599, + "loss": 3.4265692234039307, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.023382030427455902, + "learning_rate": 0.00034656246952095984, + "loss": 3.378389835357666, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.02340470254421234, + "learning_rate": 0.0003462987486341648, + "loss": 3.392937421798706, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.022166339680552483, + "learning_rate": 0.00034603499286235006, + "loss": 3.39414119720459, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.022175129503011703, + "learning_rate": 0.0003457712024180188, + "loss": 3.407447099685669, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.02368207648396492, + "learning_rate": 0.0003455073775137025, + "loss": 3.415076971054077, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.02263982594013214, + "learning_rate": 0.0003452435183619598, + "loss": 3.383986473083496, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.02147712931036949, + "learning_rate": 0.0003449796251753773, + "loss": 3.403963565826416, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.022987166419625282, + "learning_rate": 0.00034471569816656915, + "loss": 3.3883113861083984, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.02240179106593132, + "learning_rate": 0.00034445173754817646, + "loss": 3.4349911212921143, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.02206592820584774, + "learning_rate": 0.00034418774353286747, + "loss": 3.426473617553711, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.022650036960840225, + "learning_rate": 0.0003439237163333375, + "loss": 3.4251773357391357, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.023399339988827705, + "learning_rate": 0.0003436596561623084, + "loss": 3.371610641479492, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.02192695252597332, + "learning_rate": 0.0003433955632325288, + "loss": 3.387253761291504, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.02241935208439827, + "learning_rate": 0.00034313143775677353, + "loss": 3.434523105621338, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.023202672600746155, + "learning_rate": 0.00034286727994784367, + "loss": 3.369771957397461, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.021333014592528343, + "learning_rate": 0.0003426030900185665, + "loss": 3.439004421234131, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.023988714441657066, + "learning_rate": 0.0003423388681817949, + "loss": 3.396838665008545, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.023885777220129967, + "learning_rate": 0.00034207461465040793, + "loss": 3.4114322662353516, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.02263713628053665, + "learning_rate": 0.0003418103296373096, + "loss": 3.4139351844787598, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.022404303774237633, + "learning_rate": 0.00034154601335542964, + "loss": 3.4200823307037354, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.021705076098442078, + "learning_rate": 0.00034128166601772304, + "loss": 3.4092609882354736, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.02273249812424183, + "learning_rate": 0.0003410172878371695, + "loss": 3.3931846618652344, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.022552402690052986, + "learning_rate": 0.00034075287902677394, + "loss": 3.396566867828369, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.023219924420118332, + "learning_rate": 0.0003404884397995655, + "loss": 3.369025707244873, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.022560900077223778, + "learning_rate": 0.00034022397036859837, + "loss": 3.4248225688934326, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.023370858281850815, + "learning_rate": 0.0003399594709469506, + "loss": 3.431628942489624, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.02469109185039997, + "learning_rate": 0.00033969494174772465, + "loss": 3.406843662261963, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.02417670376598835, + "learning_rate": 0.0003394303829840469, + "loss": 3.445046901702881, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.0237257182598114, + "learning_rate": 0.00033916579486906764, + "loss": 3.4490597248077393, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.02295520156621933, + "learning_rate": 0.0003389011776159607, + "loss": 3.4009876251220703, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.023158475756645203, + "learning_rate": 0.0003386365314379233, + "loss": 3.4383902549743652, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.023618478327989578, + "learning_rate": 0.0003383718565481763, + "loss": 3.421168804168701, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.022055329754948616, + "learning_rate": 0.0003381071531599633, + "loss": 3.382946729660034, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.021441346034407616, + "learning_rate": 0.00033784242148655115, + "loss": 3.4464712142944336, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.023031961172819138, + "learning_rate": 0.00033757766174122934, + "loss": 3.42844295501709, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.022832952439785004, + "learning_rate": 0.00033731287413731005, + "loss": 3.398231029510498, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.022094765678048134, + "learning_rate": 0.0003370480588881278, + "loss": 3.4270074367523193, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.0218376275151968, + "learning_rate": 0.0003367832162070395, + "loss": 3.4681742191314697, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.02364739589393139, + "learning_rate": 0.000336518346307424, + "loss": 3.431920051574707, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.023402225226163864, + "learning_rate": 0.0003362534494026824, + "loss": 3.4189023971557617, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.022775836288928986, + "learning_rate": 0.0003359885257062372, + "loss": 3.409396171569824, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.02445383556187153, + "learning_rate": 0.0003357235754315328, + "loss": 3.41017746925354, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.066352367401123, + "eval_runtime": 58.1658, + "eval_samples_per_second": 41.983, + "eval_steps_per_second": 1.324, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.02710847556591034, + "learning_rate": 0.0003354585987920345, + "loss": 3.3155131340026855, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.03428799286484718, + "learning_rate": 0.0003351935960012296, + "loss": 3.327408790588379, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.03434379771351814, + "learning_rate": 0.0003349285672726259, + "loss": 3.3282251358032227, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.03405233472585678, + "learning_rate": 0.0003346635128197522, + "loss": 3.2932448387145996, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.031194346025586128, + "learning_rate": 0.0003343984328561581, + "loss": 3.336615800857544, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.031816478818655014, + "learning_rate": 0.00033413332759541376, + "loss": 3.3375275135040283, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.03413299471139908, + "learning_rate": 0.0003338681972511098, + "loss": 3.3710694313049316, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.034200165420770645, + "learning_rate": 0.0003336030420368568, + "loss": 3.3248214721679688, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.030694160610437393, + "learning_rate": 0.0003333378621662857, + "loss": 3.3297901153564453, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.031172804534435272, + "learning_rate": 0.00033307265785304684, + "loss": 3.287217140197754, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.031914614140987396, + "learning_rate": 0.00033280742931081076, + "loss": 3.3218531608581543, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.0332055389881134, + "learning_rate": 0.00033254217675326737, + "loss": 3.3454785346984863, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.034956980496644974, + "learning_rate": 0.0003322769003941257, + "loss": 3.3193702697753906, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.03618055582046509, + "learning_rate": 0.00033201160044711423, + "loss": 3.3661017417907715, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.030322827398777008, + "learning_rate": 0.0003317462771259802, + "loss": 3.353989601135254, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.029768960550427437, + "learning_rate": 0.00033148093064449006, + "loss": 3.3254497051239014, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.028036782518029213, + "learning_rate": 0.0003312155612164284, + "loss": 3.3276920318603516, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.028774483129382133, + "learning_rate": 0.00033095016905559883, + "loss": 3.3458826541900635, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.030403515323996544, + "learning_rate": 0.0003306847543758227, + "loss": 3.3332793712615967, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.028468478471040726, + "learning_rate": 0.0003304193173909401, + "loss": 3.3167755603790283, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.029126476496458054, + "learning_rate": 0.00033015385831480873, + "loss": 3.299426794052124, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.030746638774871826, + "learning_rate": 0.0003298883773613043, + "loss": 3.3581809997558594, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.030241351574659348, + "learning_rate": 0.0003296228747443197, + "loss": 3.3415942192077637, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.029754415154457092, + "learning_rate": 0.0003293573506777659, + "loss": 3.3525662422180176, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.026896746829152107, + "learning_rate": 0.00032909180537557076, + "loss": 3.321808099746704, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.026361066848039627, + "learning_rate": 0.00032882623905167917, + "loss": 3.329258441925049, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.027747981250286102, + "learning_rate": 0.00032856065192005335, + "loss": 3.3774523735046387, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.026270100846886635, + "learning_rate": 0.00032829504419467194, + "loss": 3.3419976234436035, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.02731695957481861, + "learning_rate": 0.0003280294160895303, + "loss": 3.3068814277648926, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.026583299040794373, + "learning_rate": 0.0003277637678186402, + "loss": 3.303804874420166, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.026615949347615242, + "learning_rate": 0.00032749809959602973, + "loss": 3.3812599182128906, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.027221621945500374, + "learning_rate": 0.0003272324116357428, + "loss": 3.327211618423462, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.025025667622685432, + "learning_rate": 0.0003269667041518395, + "loss": 3.339163303375244, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.02593836560845375, + "learning_rate": 0.0003267009773583956, + "loss": 3.3517372608184814, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.023697076365351677, + "learning_rate": 0.00032643523146950234, + "loss": 3.379819869995117, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.02425902709364891, + "learning_rate": 0.0003261694666992664, + "loss": 3.390653610229492, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.025133250281214714, + "learning_rate": 0.0003259036832618096, + "loss": 3.364495277404785, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.025542905554175377, + "learning_rate": 0.0003256378813712688, + "loss": 3.3584814071655273, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.026041125878691673, + "learning_rate": 0.000325372061241796, + "loss": 3.36777400970459, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.02391037344932556, + "learning_rate": 0.00032510622308755746, + "loss": 3.3809385299682617, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.02379394695162773, + "learning_rate": 0.00032484036712273424, + "loss": 3.354257583618164, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.0234722550958395, + "learning_rate": 0.0003245744935615219, + "loss": 3.3736205101013184, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.024123650044202805, + "learning_rate": 0.0003243086026181296, + "loss": 3.3589959144592285, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.024013696238398552, + "learning_rate": 0.00032404269450678116, + "loss": 3.353874444961548, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.022617843002080917, + "learning_rate": 0.00032377676944171375, + "loss": 3.364056348800659, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.022730547934770584, + "learning_rate": 0.00032351082763717857, + "loss": 3.378067970275879, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.022257911041378975, + "learning_rate": 0.0003232448693074399, + "loss": 3.379342555999756, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.023247094824910164, + "learning_rate": 0.00032297889466677575, + "loss": 3.3537116050720215, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.022764021530747414, + "learning_rate": 0.000322712903929477, + "loss": 3.3969063758850098, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.02260403335094452, + "learning_rate": 0.0003224468973098477, + "loss": 3.3727574348449707, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.021772351115942, + "learning_rate": 0.0003221808750222044, + "loss": 3.335434675216675, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.02238459885120392, + "learning_rate": 0.0003219148372808766, + "loss": 3.3440277576446533, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.0230540968477726, + "learning_rate": 0.00032164878430020606, + "loss": 3.3997583389282227, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.0220821350812912, + "learning_rate": 0.00032138271629454684, + "loss": 3.377815008163452, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.021497417241334915, + "learning_rate": 0.00032111663347826505, + "loss": 3.3696370124816895, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.02163885347545147, + "learning_rate": 0.00032085053606573896, + "loss": 3.3489794731140137, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.022394269704818726, + "learning_rate": 0.0003205844242713584, + "loss": 3.3792812824249268, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.022390708327293396, + "learning_rate": 0.0003203182983095248, + "loss": 3.3733272552490234, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.023712158203125, + "learning_rate": 0.0003200521583946511, + "loss": 3.3716251850128174, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.02398860827088356, + "learning_rate": 0.0003197860047411613, + "loss": 3.381779670715332, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.02309161052107811, + "learning_rate": 0.00031951983756349066, + "loss": 3.3923277854919434, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.021400388330221176, + "learning_rate": 0.00031925365707608536, + "loss": 3.407397747039795, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.022534485906362534, + "learning_rate": 0.00031898746349340204, + "loss": 3.3740391731262207, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.023533202707767487, + "learning_rate": 0.0003187212570299082, + "loss": 3.3877944946289062, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.023398835211992264, + "learning_rate": 0.0003184550379000815, + "loss": 3.3936471939086914, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.020927201956510544, + "learning_rate": 0.00031818880631841005, + "loss": 3.339545249938965, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.021453995257616043, + "learning_rate": 0.0003179225624993915, + "loss": 3.3647522926330566, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.022847114130854607, + "learning_rate": 0.0003176563066575341, + "loss": 3.405144691467285, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.022766996175050735, + "learning_rate": 0.0003173900390073549, + "loss": 3.392810344696045, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.022103948518633842, + "learning_rate": 0.0003171237597633813, + "loss": 3.3973004817962646, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.022282181307673454, + "learning_rate": 0.0003168574691401496, + "loss": 3.409998893737793, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.025435522198677063, + "learning_rate": 0.0003165911673522053, + "loss": 3.423318386077881, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.078423500061035, + "eval_runtime": 57.3189, + "eval_samples_per_second": 42.604, + "eval_steps_per_second": 1.343, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.02624760940670967, + "learning_rate": 0.00031632485461410294, + "loss": 3.261288642883301, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.03552987053990364, + "learning_rate": 0.00031605853114040583, + "loss": 3.2682106494903564, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.040380869060754776, + "learning_rate": 0.00031579219714568613, + "loss": 3.2813782691955566, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.03861293941736221, + "learning_rate": 0.0003155258528445242, + "loss": 3.258193016052246, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.04091731086373329, + "learning_rate": 0.0003152594984515089, + "loss": 3.3052220344543457, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.043269723653793335, + "learning_rate": 0.00031499313418123704, + "loss": 3.2820217609405518, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.04580320790410042, + "learning_rate": 0.0003147267602483138, + "loss": 3.312005043029785, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.047336459159851074, + "learning_rate": 0.00031446037686735153, + "loss": 3.3217713832855225, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.04587360844016075, + "learning_rate": 0.0003141939842529706, + "loss": 3.2869949340820312, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.04002532362937927, + "learning_rate": 0.0003139275826197987, + "loss": 3.2879223823547363, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.0368693508207798, + "learning_rate": 0.00031366117218247094, + "loss": 3.2983577251434326, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.037317775189876556, + "learning_rate": 0.00031339475315562916, + "loss": 3.3062524795532227, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.03655700013041496, + "learning_rate": 0.00031312832575392267, + "loss": 3.31072735786438, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.033466216176748276, + "learning_rate": 0.0003128618901920071, + "loss": 3.2954514026641846, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.030785970389842987, + "learning_rate": 0.0003125954466845446, + "loss": 3.27952241897583, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.029669342562556267, + "learning_rate": 0.00031232899544620404, + "loss": 3.2613229751586914, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.033505845814943314, + "learning_rate": 0.00031206253669166045, + "loss": 3.2987258434295654, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.032395437359809875, + "learning_rate": 0.0003117960706355947, + "loss": 3.296708583831787, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.031322430819272995, + "learning_rate": 0.0003115295974926936, + "loss": 3.3160252571105957, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.029345186427235603, + "learning_rate": 0.00031126311747765005, + "loss": 3.307143211364746, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.026715675368905067, + "learning_rate": 0.00031099663080516186, + "loss": 3.286012649536133, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.028592750430107117, + "learning_rate": 0.0003107301376899327, + "loss": 3.2528645992279053, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.028861764818429947, + "learning_rate": 0.00031046363834667123, + "loss": 3.2795748710632324, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.02609528973698616, + "learning_rate": 0.00031019713299009123, + "loss": 3.326317071914673, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.027718454599380493, + "learning_rate": 0.000309930621834911, + "loss": 3.268110513687134, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.026382867246866226, + "learning_rate": 0.0003096641050958541, + "loss": 3.294499397277832, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.027083583176136017, + "learning_rate": 0.00030939758298764795, + "loss": 3.3202130794525146, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.02516467310488224, + "learning_rate": 0.00030913105572502483, + "loss": 3.3485021591186523, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.024148069322109222, + "learning_rate": 0.00030886452352272095, + "loss": 3.3099541664123535, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.024552857503294945, + "learning_rate": 0.00030859798659547624, + "loss": 3.2885942459106445, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.02551114559173584, + "learning_rate": 0.0003083314451580349, + "loss": 3.3232040405273438, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.025423647835850716, + "learning_rate": 0.0003080648994251445, + "loss": 3.3097832202911377, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.023211417719721794, + "learning_rate": 0.0003077983496115561, + "loss": 3.3075966835021973, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.024623898789286613, + "learning_rate": 0.00030753179593202406, + "loss": 3.2917377948760986, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.02677663415670395, + "learning_rate": 0.0003072652386013059, + "loss": 3.356265068054199, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.02673642337322235, + "learning_rate": 0.00030699867783416207, + "loss": 3.314903736114502, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.024466225877404213, + "learning_rate": 0.0003067321138453557, + "loss": 3.30912446975708, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.025370748713612556, + "learning_rate": 0.0003064655468496527, + "loss": 3.363640069961548, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.026536114513874054, + "learning_rate": 0.0003061989770618211, + "loss": 3.304851531982422, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.026536893099546432, + "learning_rate": 0.00030593240469663154, + "loss": 3.342902898788452, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.02645307406783104, + "learning_rate": 0.0003056658299688563, + "loss": 3.3164432048797607, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.02523060515522957, + "learning_rate": 0.00030539925309327017, + "loss": 3.34376859664917, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.02590431459248066, + "learning_rate": 0.0003051326742846491, + "loss": 3.3479061126708984, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.0230655986815691, + "learning_rate": 0.000304866093757771, + "loss": 3.2890982627868652, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.02403893694281578, + "learning_rate": 0.00030459951172741474, + "loss": 3.345456600189209, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.02238907851278782, + "learning_rate": 0.0003043329284083609, + "loss": 3.3471767902374268, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.021377569064497948, + "learning_rate": 0.00030406634401539053, + "loss": 3.319232225418091, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.02272353693842888, + "learning_rate": 0.00030379975876328614, + "loss": 3.316892623901367, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.02413824200630188, + "learning_rate": 0.00030353317286683053, + "loss": 3.3450968265533447, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.02334682270884514, + "learning_rate": 0.0003032665865408073, + "loss": 3.3640642166137695, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.021966520696878433, + "learning_rate": 0.000303, + "loss": 3.358555316925049, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.023984894156455994, + "learning_rate": 0.0003027334134591926, + "loss": 3.358398914337158, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.023520268499851227, + "learning_rate": 0.00030246682713316945, + "loss": 3.3224587440490723, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.0231285709887743, + "learning_rate": 0.0003022002412367138, + "loss": 3.3671562671661377, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.022818170487880707, + "learning_rate": 0.00030193365598460946, + "loss": 3.361027717590332, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.02280694991350174, + "learning_rate": 0.0003016670715916391, + "loss": 3.2923383712768555, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.021362345665693283, + "learning_rate": 0.00030140048827258524, + "loss": 3.3346734046936035, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.02372218668460846, + "learning_rate": 0.0003011339062422289, + "loss": 3.354748249053955, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.022949764505028725, + "learning_rate": 0.00030086732571535076, + "loss": 3.3857789039611816, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.021954255178570747, + "learning_rate": 0.0003006007469067297, + "loss": 3.343330144882202, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.022438595071434975, + "learning_rate": 0.0003003341700311436, + "loss": 3.328618049621582, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.021969925612211227, + "learning_rate": 0.0003000675953033685, + "loss": 3.355971574783325, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.022706981748342514, + "learning_rate": 0.0002998010229381789, + "loss": 3.3529577255249023, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.024083172902464867, + "learning_rate": 0.0002995344531503474, + "loss": 3.348170757293701, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.023259354755282402, + "learning_rate": 0.0002992678861546442, + "loss": 3.3474369049072266, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.02218157798051834, + "learning_rate": 0.00029900132216583786, + "loss": 3.369053840637207, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.02305331453680992, + "learning_rate": 0.0002987347613986941, + "loss": 3.3622641563415527, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.022796178236603737, + "learning_rate": 0.0002984682040679759, + "loss": 3.351510524749756, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.021299205720424652, + "learning_rate": 0.00029820165038844395, + "loss": 3.339406728744507, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.022786924615502357, + "learning_rate": 0.00029793510057485556, + "loss": 3.3182196617126465, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.021729128435254097, + "learning_rate": 0.00029766855484196513, + "loss": 3.383904218673706, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.024604283273220062, + "learning_rate": 0.0002974020134045238, + "loss": 3.2912864685058594, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.0894880294799805, + "eval_runtime": 58.3967, + "eval_samples_per_second": 41.817, + "eval_steps_per_second": 1.319, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.02673362009227276, + "learning_rate": 0.00029713547647727904, + "loss": 3.270986795425415, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.04036599025130272, + "learning_rate": 0.00029686894427497515, + "loss": 3.2358999252319336, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.047250088304281235, + "learning_rate": 0.0002966024170123519, + "loss": 3.2706189155578613, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.04642819985747337, + "learning_rate": 0.0002963358949041459, + "loss": 3.236340284347534, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.039405327290296555, + "learning_rate": 0.00029606937816508897, + "loss": 3.203136682510376, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.038748081773519516, + "learning_rate": 0.00029580286700990887, + "loss": 3.2636070251464844, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.039233580231666565, + "learning_rate": 0.00029553636165332876, + "loss": 3.2751896381378174, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.03429320082068443, + "learning_rate": 0.0002952698623100673, + "loss": 3.2765679359436035, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.03267038241028786, + "learning_rate": 0.0002950033691948382, + "loss": 3.2387681007385254, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.03349406272172928, + "learning_rate": 0.00029473688252234994, + "loss": 3.282162666320801, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.032002877444028854, + "learning_rate": 0.0002944704025073064, + "loss": 3.246623992919922, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.031509071588516235, + "learning_rate": 0.00029420392936440525, + "loss": 3.2374932765960693, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.02958357334136963, + "learning_rate": 0.00029393746330833953, + "loss": 3.24362850189209, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.03248635679483414, + "learning_rate": 0.0002936710045537959, + "loss": 3.237210273742676, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.03230655938386917, + "learning_rate": 0.0002934045533154554, + "loss": 3.251129150390625, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.029688173905014992, + "learning_rate": 0.00029313810980799296, + "loss": 3.2141575813293457, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.027773940935730934, + "learning_rate": 0.00029287167424607726, + "loss": 3.264249086380005, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.02687031403183937, + "learning_rate": 0.00029260524684437077, + "loss": 3.2513632774353027, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.02503623627126217, + "learning_rate": 0.00029233882781752905, + "loss": 3.217961072921753, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.027071228250861168, + "learning_rate": 0.00029207241738020136, + "loss": 3.2784645557403564, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.02600773423910141, + "learning_rate": 0.0002918060157470294, + "loss": 3.2641382217407227, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.026998162269592285, + "learning_rate": 0.00029153962313264856, + "loss": 3.2646782398223877, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.025579402223229408, + "learning_rate": 0.0002912732397516862, + "loss": 3.2700581550598145, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.02645333670079708, + "learning_rate": 0.00029100686581876283, + "loss": 3.2545266151428223, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.02717350795865059, + "learning_rate": 0.0002907405015484911, + "loss": 3.2795705795288086, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.02675914391875267, + "learning_rate": 0.0002904741471554758, + "loss": 3.2936904430389404, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.02751768007874489, + "learning_rate": 0.0002902078028543139, + "loss": 3.2329440116882324, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.025446882471442223, + "learning_rate": 0.00028994146885959416, + "loss": 3.274294376373291, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.02365768514573574, + "learning_rate": 0.00028967514538589715, + "loss": 3.2585229873657227, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.0266884732991457, + "learning_rate": 0.0002894088326477947, + "loss": 3.2872962951660156, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.02572469227015972, + "learning_rate": 0.0002891425308598503, + "loss": 3.253828287124634, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.02540273219347, + "learning_rate": 0.00028887624023661866, + "loss": 3.303926944732666, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.024694625288248062, + "learning_rate": 0.00028860996099264505, + "loss": 3.264145851135254, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.027233170345425606, + "learning_rate": 0.000288343693342466, + "loss": 3.286924362182617, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02635285072028637, + "learning_rate": 0.00028807743750060847, + "loss": 3.2930331230163574, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.025157805532217026, + "learning_rate": 0.00028781119368159004, + "loss": 3.2906906604766846, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.024729596450924873, + "learning_rate": 0.00028754496209991843, + "loss": 3.277128219604492, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.023809000849723816, + "learning_rate": 0.0002872787429700917, + "loss": 3.2957217693328857, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02474421076476574, + "learning_rate": 0.0002870125365065979, + "loss": 3.27791690826416, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.02439294196665287, + "learning_rate": 0.00028674634292391463, + "loss": 3.2912981510162354, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.023601334542036057, + "learning_rate": 0.0002864801624365093, + "loss": 3.317446231842041, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.0253264419734478, + "learning_rate": 0.0002862139952588387, + "loss": 3.2742302417755127, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.02389134280383587, + "learning_rate": 0.000285947841605349, + "loss": 3.281032085418701, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.02451813407242298, + "learning_rate": 0.00028568170169047523, + "loss": 3.3339295387268066, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.024324845522642136, + "learning_rate": 0.00028541557572864154, + "loss": 3.290402889251709, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.024939008057117462, + "learning_rate": 0.000285149463934261, + "loss": 3.2933621406555176, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.024450505152344704, + "learning_rate": 0.0002848833665217349, + "loss": 3.3055014610290527, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.02262270078063011, + "learning_rate": 0.0002846172837054532, + "loss": 3.2964205741882324, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.02311946265399456, + "learning_rate": 0.0002843512156997939, + "loss": 3.3102402687072754, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.024362705647945404, + "learning_rate": 0.0002840851627191234, + "loss": 3.3020925521850586, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.024785032495856285, + "learning_rate": 0.0002838191249777955, + "loss": 3.287290096282959, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.02358846180140972, + "learning_rate": 0.00028355310269015225, + "loss": 3.334261894226074, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.024330584332346916, + "learning_rate": 0.00028328709607052297, + "loss": 3.2799081802368164, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.023546984419226646, + "learning_rate": 0.0002830211053332242, + "loss": 3.3228847980499268, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.023060832172632217, + "learning_rate": 0.0002827551306925601, + "loss": 3.325793981552124, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.02388455718755722, + "learning_rate": 0.00028248917236282147, + "loss": 3.2892062664031982, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.023740701377391815, + "learning_rate": 0.0002822232305582863, + "loss": 3.2969369888305664, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.02458217367529869, + "learning_rate": 0.0002819573054932188, + "loss": 3.315610408782959, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.023173579946160316, + "learning_rate": 0.0002816913973818703, + "loss": 3.2561936378479004, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.024137936532497406, + "learning_rate": 0.0002814255064384781, + "loss": 3.2839202880859375, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.022436272352933884, + "learning_rate": 0.00028115963287726563, + "loss": 3.3057351112365723, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.023997878655791283, + "learning_rate": 0.0002808937769124425, + "loss": 3.3583154678344727, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.024119896814227104, + "learning_rate": 0.000280627938758204, + "loss": 3.322892427444458, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.022283365949988365, + "learning_rate": 0.00028036211862873117, + "loss": 3.3145735263824463, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.023433363065123558, + "learning_rate": 0.0002800963167381904, + "loss": 3.3564529418945312, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.02224746160209179, + "learning_rate": 0.00027983053330073355, + "loss": 3.2985916137695312, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.023786570876836777, + "learning_rate": 0.00027956476853049765, + "loss": 3.302539587020874, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.025200961157679558, + "learning_rate": 0.00027929902264160435, + "loss": 3.309976577758789, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.023849809542298317, + "learning_rate": 0.00027903329584816046, + "loss": 3.343660354614258, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.023311031982302666, + "learning_rate": 0.0002787675883642572, + "loss": 3.2978909015655518, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.023174891248345375, + "learning_rate": 0.00027850190040397036, + "loss": 3.326188802719116, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.02681034244596958, + "learning_rate": 0.00027823623218135976, + "loss": 3.3422017097473145, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.109009265899658, + "eval_runtime": 58.0506, + "eval_samples_per_second": 42.067, + "eval_steps_per_second": 1.326, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.028057295829057693, + "learning_rate": 0.0002779705839104697, + "loss": 3.1825037002563477, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.034786805510520935, + "learning_rate": 0.0002777049558053281, + "loss": 3.200389862060547, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.03751441463828087, + "learning_rate": 0.0002774393480799466, + "loss": 3.2047512531280518, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.035590436309576035, + "learning_rate": 0.00027717376094832076, + "loss": 3.1735124588012695, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.03426426649093628, + "learning_rate": 0.0002769081946244293, + "loss": 3.199483871459961, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.037030890583992004, + "learning_rate": 0.00027664264932223407, + "loss": 3.213322639465332, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.032468877732753754, + "learning_rate": 0.00027637712525568024, + "loss": 3.179354190826416, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.027202758938074112, + "learning_rate": 0.0002761116226386958, + "loss": 3.216036319732666, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.031184563413262367, + "learning_rate": 0.00027584614168519125, + "loss": 3.208205223083496, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.029414812102913857, + "learning_rate": 0.00027558068260905976, + "loss": 3.180370569229126, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.02679901383817196, + "learning_rate": 0.00027531524562417734, + "loss": 3.210325002670288, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.02863631211221218, + "learning_rate": 0.00027504983094440126, + "loss": 3.1601076126098633, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.029135366901755333, + "learning_rate": 0.0002747844387835716, + "loss": 3.2262887954711914, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.030535712838172913, + "learning_rate": 0.0002745190693555099, + "loss": 3.23160719871521, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.030563877895474434, + "learning_rate": 0.0002742537228740197, + "loss": 3.255211353302002, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.02906505949795246, + "learning_rate": 0.00027398839955288575, + "loss": 3.2006797790527344, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.030871668830513954, + "learning_rate": 0.00027372309960587424, + "loss": 3.2442867755889893, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.02913421392440796, + "learning_rate": 0.0002734578232467327, + "loss": 3.218442440032959, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.030087953433394432, + "learning_rate": 0.0002731925706891891, + "loss": 3.2164535522460938, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.034147944301366806, + "learning_rate": 0.0002729273421469532, + "loss": 3.218660831451416, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.03593966364860535, + "learning_rate": 0.00027266213783371434, + "loss": 3.2240614891052246, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.032090652734041214, + "learning_rate": 0.00027239695796314327, + "loss": 3.2309441566467285, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.02776424027979374, + "learning_rate": 0.0002721318027488902, + "loss": 3.2702016830444336, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.027174271643161774, + "learning_rate": 0.00027186667240458617, + "loss": 3.19810152053833, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.028479408472776413, + "learning_rate": 0.0002716015671438419, + "loss": 3.2178421020507812, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.026820644736289978, + "learning_rate": 0.0002713364871802478, + "loss": 3.2634482383728027, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.026427917182445526, + "learning_rate": 0.0002710714327273741, + "loss": 3.2452914714813232, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.02679610252380371, + "learning_rate": 0.0002708064039987704, + "loss": 3.243825912475586, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.026115626096725464, + "learning_rate": 0.00027054140120796546, + "loss": 3.218787670135498, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.026686571538448334, + "learning_rate": 0.0002702764245684672, + "loss": 3.2785303592681885, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.02685859426856041, + "learning_rate": 0.0002700114742937627, + "loss": 3.2572226524353027, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.026968171820044518, + "learning_rate": 0.00026974655059731754, + "loss": 3.2625553607940674, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.025397246703505516, + "learning_rate": 0.0002694816536925759, + "loss": 3.262303352355957, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.025660471990704536, + "learning_rate": 0.00026921678379296057, + "loss": 3.269401788711548, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.02713891491293907, + "learning_rate": 0.0002689519411118722, + "loss": 3.2416529655456543, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.027591092512011528, + "learning_rate": 0.00026868712586269, + "loss": 3.2814316749572754, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.024462848901748657, + "learning_rate": 0.00026842233825877064, + "loss": 3.249725103378296, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.024868478998541832, + "learning_rate": 0.0002681575785134487, + "loss": 3.231574535369873, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.027015401050448418, + "learning_rate": 0.0002678928468400366, + "loss": 3.282627582550049, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.02678140625357628, + "learning_rate": 0.00026762814345182367, + "loss": 3.2208900451660156, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.026830527931451797, + "learning_rate": 0.0002673634685620767, + "loss": 3.260575294494629, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.02430890128016472, + "learning_rate": 0.0002670988223840393, + "loss": 3.2522997856140137, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.027047613635659218, + "learning_rate": 0.0002668342051309324, + "loss": 3.2721078395843506, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.026861369609832764, + "learning_rate": 0.00026656961701595306, + "loss": 3.2235946655273438, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.026650549843907356, + "learning_rate": 0.0002663050582522753, + "loss": 3.302743434906006, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.02551819011569023, + "learning_rate": 0.00026604052905304946, + "loss": 3.265763282775879, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.024843450635671616, + "learning_rate": 0.00026577602963140156, + "loss": 3.272397518157959, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.024423206225037575, + "learning_rate": 0.0002655115602004345, + "loss": 3.253495454788208, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.024180661886930466, + "learning_rate": 0.00026524712097322604, + "loss": 3.260438919067383, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.023554982617497444, + "learning_rate": 0.0002649827121628305, + "loss": 3.2342851161956787, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.023470086976885796, + "learning_rate": 0.00026471833398227694, + "loss": 3.251622438430786, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.024405231699347496, + "learning_rate": 0.0002644539866445703, + "loss": 3.2588815689086914, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.024147814139723778, + "learning_rate": 0.00026418967036269045, + "loss": 3.2723565101623535, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.023217910900712013, + "learning_rate": 0.0002639253853495921, + "loss": 3.2555792331695557, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.024921158328652382, + "learning_rate": 0.00026366113181820506, + "loss": 3.26884126663208, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.02455919235944748, + "learning_rate": 0.00026339690998143346, + "loss": 3.2674927711486816, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.024725932627916336, + "learning_rate": 0.00026313272005215637, + "loss": 3.2511110305786133, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.025769414380192757, + "learning_rate": 0.00026286856224322645, + "loss": 3.290651321411133, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.025928806513547897, + "learning_rate": 0.00026260443676747114, + "loss": 3.302396297454834, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.02499072253704071, + "learning_rate": 0.00026234034383769157, + "loss": 3.2813124656677246, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.025431277230381966, + "learning_rate": 0.0002620762836666625, + "loss": 3.234684705734253, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.026871860027313232, + "learning_rate": 0.00026181225646713257, + "loss": 3.2800331115722656, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.02578776702284813, + "learning_rate": 0.0002615482624518236, + "loss": 3.3117146492004395, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.026318378746509552, + "learning_rate": 0.0002612843018334309, + "loss": 3.257615089416504, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.025542527437210083, + "learning_rate": 0.00026102037482462256, + "loss": 3.2969868183135986, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.025723429396748543, + "learning_rate": 0.0002607564816380403, + "loss": 3.290259838104248, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.024588774889707565, + "learning_rate": 0.0002604926224862975, + "loss": 3.284090280532837, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.024619122967123985, + "learning_rate": 0.0002602287975819811, + "loss": 3.284879207611084, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.024441618472337723, + "learning_rate": 0.00025996500713765, + "loss": 3.283104658126831, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.02402760274708271, + "learning_rate": 0.0002597012513658352, + "loss": 3.3013617992401123, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.025028521195054054, + "learning_rate": 0.00025943753047904026, + "loss": 3.2892568111419678, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.029445001855492592, + "learning_rate": 0.0002591738446897401, + "loss": 3.2474446296691895, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.128176689147949, + "eval_runtime": 58.8588, + "eval_samples_per_second": 41.489, + "eval_steps_per_second": 1.308, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.0298826452344656, + "learning_rate": 0.00025891019421038177, + "loss": 3.138103485107422, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.0387548953294754, + "learning_rate": 0.00025864657925338345, + "loss": 3.161492347717285, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.03816070780158043, + "learning_rate": 0.00025838300003113514, + "loss": 3.1739673614501953, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.03409738838672638, + "learning_rate": 0.0002581194567559979, + "loss": 3.182478189468384, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.033135563135147095, + "learning_rate": 0.00025785594964030365, + "loss": 3.1720199584960938, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.03203624114394188, + "learning_rate": 0.0002575924788963554, + "loss": 3.1699862480163574, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.031237754970788956, + "learning_rate": 0.0002573290447364268, + "loss": 3.179363250732422, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.03342146426439285, + "learning_rate": 0.0002570656473727619, + "loss": 3.1822280883789062, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.03561089560389519, + "learning_rate": 0.0002568022870175752, + "loss": 3.1767094135284424, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.034710463136434555, + "learning_rate": 0.00025653896388305134, + "loss": 3.218817949295044, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.030870331451296806, + "learning_rate": 0.0002562756781813453, + "loss": 3.168602705001831, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.028004391118884087, + "learning_rate": 0.00025601243012458126, + "loss": 3.224155902862549, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.030953701585531235, + "learning_rate": 0.00025574921992485366, + "loss": 3.1789557933807373, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.030289432033896446, + "learning_rate": 0.000255486047794226, + "loss": 3.1890387535095215, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.030684132128953934, + "learning_rate": 0.0002552229139447316, + "loss": 3.195427417755127, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.03022196516394615, + "learning_rate": 0.00025495981858837235, + "loss": 3.1984965801239014, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.026935536414384842, + "learning_rate": 0.0002546967619371195, + "loss": 3.2014272212982178, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.02702922560274601, + "learning_rate": 0.0002544337442029133, + "loss": 3.205538034439087, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.026682600378990173, + "learning_rate": 0.00025417076559766194, + "loss": 3.1954591274261475, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.0273384228348732, + "learning_rate": 0.000253907826333243, + "loss": 3.1782338619232178, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.028587577864527702, + "learning_rate": 0.00025364492662150145, + "loss": 3.199565887451172, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.028848132118582726, + "learning_rate": 0.0002533820666742514, + "loss": 3.181786298751831, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.027535341680049896, + "learning_rate": 0.00025311924670327364, + "loss": 3.2013628482818604, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.02756025455892086, + "learning_rate": 0.00025285646692031804, + "loss": 3.1749587059020996, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.027358995750546455, + "learning_rate": 0.0002525937275371013, + "loss": 3.161881923675537, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.025918981060385704, + "learning_rate": 0.00025233102876530785, + "loss": 3.225224256515503, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.026479976251721382, + "learning_rate": 0.0002520683708165894, + "loss": 3.2129316329956055, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.02685079537332058, + "learning_rate": 0.0002518057539025646, + "loss": 3.1651554107666016, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.027378706261515617, + "learning_rate": 0.00025154317823481917, + "loss": 3.193108558654785, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.027240974828600883, + "learning_rate": 0.0002512806440249056, + "loss": 3.190605640411377, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.028065402060747147, + "learning_rate": 0.00025101815148434296, + "loss": 3.2068259716033936, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.028329014778137207, + "learning_rate": 0.00025075570082461684, + "loss": 3.2316060066223145, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.028920065611600876, + "learning_rate": 0.0002504932922571789, + "loss": 3.1933932304382324, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.028299354016780853, + "learning_rate": 0.00025023092599344706, + "loss": 3.192711353302002, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.025996020063757896, + "learning_rate": 0.0002499686022448049, + "loss": 3.243171453475952, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.0264595914632082, + "learning_rate": 0.00024970632122260216, + "loss": 3.199758529663086, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.02704528719186783, + "learning_rate": 0.0002494440831381538, + "loss": 3.2390713691711426, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.026571933180093765, + "learning_rate": 0.00024918188820274016, + "loss": 3.219440460205078, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.025562983006238937, + "learning_rate": 0.0002489197366276073, + "loss": 3.2538228034973145, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.024683184921741486, + "learning_rate": 0.0002486576286239657, + "loss": 3.2372591495513916, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.025656528770923615, + "learning_rate": 0.0002483955644029913, + "loss": 3.250558853149414, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.024807540699839592, + "learning_rate": 0.0002481335441758243, + "loss": 3.1856563091278076, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.026149194687604904, + "learning_rate": 0.0002478715681535699, + "loss": 3.172186851501465, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.02617705799639225, + "learning_rate": 0.00024760963654729717, + "loss": 3.250445604324341, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.026542512699961662, + "learning_rate": 0.0002473477495680397, + "loss": 3.220273494720459, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.024963509291410446, + "learning_rate": 0.00024708590742679546, + "loss": 3.2330946922302246, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.027144264429807663, + "learning_rate": 0.00024682411033452546, + "loss": 3.1860971450805664, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.02665683813393116, + "learning_rate": 0.0002465623585021553, + "loss": 3.257209062576294, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.024872658774256706, + "learning_rate": 0.0002463006521405733, + "loss": 3.2266182899475098, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.025235185399651527, + "learning_rate": 0.0002460389914606319, + "loss": 3.2134504318237305, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.025029495358467102, + "learning_rate": 0.00024577737667314615, + "loss": 3.2150087356567383, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.025630896911025047, + "learning_rate": 0.00024551580798889446, + "loss": 3.282879114151001, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.02722341939806938, + "learning_rate": 0.00024525428561861804, + "loss": 3.2366206645965576, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.02522372268140316, + "learning_rate": 0.0002449928097730207, + "loss": 3.2840523719787598, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.02494790591299534, + "learning_rate": 0.00024473138066276876, + "loss": 3.2405319213867188, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.026662737131118774, + "learning_rate": 0.0002444699984984909, + "loss": 3.239692211151123, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.026082608848810196, + "learning_rate": 0.0002442086634907782, + "loss": 3.2156195640563965, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.02423974685370922, + "learning_rate": 0.00024394737585018328, + "loss": 3.1863760948181152, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.025226688012480736, + "learning_rate": 0.00024368613578722113, + "loss": 3.20874285697937, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.02428225241601467, + "learning_rate": 0.00024342494351236799, + "loss": 3.2209601402282715, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.023819707334041595, + "learning_rate": 0.0002431637992360618, + "loss": 3.2564258575439453, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.025661710649728775, + "learning_rate": 0.00024290270316870184, + "loss": 3.244133949279785, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.02423049882054329, + "learning_rate": 0.0002426416555206484, + "loss": 3.2376630306243896, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.02382473088800907, + "learning_rate": 0.000242380656502223, + "loss": 3.2951290607452393, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.02488875202834606, + "learning_rate": 0.00024211970632370756, + "loss": 3.2848310470581055, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.02575223706662655, + "learning_rate": 0.0002418588051953453, + "loss": 3.2574503421783447, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.02462848648428917, + "learning_rate": 0.0002415979533273392, + "loss": 3.2464442253112793, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.024912388995289803, + "learning_rate": 0.0002413371509298531, + "loss": 3.2459192276000977, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.025020884349942207, + "learning_rate": 0.00024107639821301078, + "loss": 3.213622808456421, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.026330886408686638, + "learning_rate": 0.00024081569538689587, + "loss": 3.2305431365966797, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.02708602324128151, + "learning_rate": 0.0002405550426615521, + "loss": 3.2487688064575195, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.030908891931176186, + "learning_rate": 0.00024029444024698248, + "loss": 3.2110159397125244, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.146673202514648, + "eval_runtime": 58.7379, + "eval_samples_per_second": 41.574, + "eval_steps_per_second": 1.311, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.02919941022992134, + "learning_rate": 0.00024003388835314981, + "loss": 3.1533267498016357, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.03379782289266586, + "learning_rate": 0.00023977338718997595, + "loss": 3.1436538696289062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.033698830753564835, + "learning_rate": 0.00023951293696734195, + "loss": 3.15297532081604, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.03077736124396324, + "learning_rate": 0.00023925253789508806, + "loss": 3.128848075866699, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.031112324446439743, + "learning_rate": 0.00023899219018301286, + "loss": 3.1461119651794434, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.03346852958202362, + "learning_rate": 0.00023873189404087405, + "loss": 3.113943099975586, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.03307225927710533, + "learning_rate": 0.00023847164967838726, + "loss": 3.147498607635498, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.030230754986405373, + "learning_rate": 0.00023821145730522692, + "loss": 3.1162924766540527, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.02982981689274311, + "learning_rate": 0.0002379513171310251, + "loss": 3.144481658935547, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.031735632568597794, + "learning_rate": 0.00023769122936537207, + "loss": 3.133862018585205, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.0312766470015049, + "learning_rate": 0.00023743119421781614, + "loss": 3.110287666320801, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.029902562499046326, + "learning_rate": 0.00023717121189786256, + "loss": 3.1384904384613037, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.0306564774364233, + "learning_rate": 0.00023691128261497463, + "loss": 3.1138200759887695, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.03084711916744709, + "learning_rate": 0.0002366514065785725, + "loss": 3.155240535736084, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.03183220326900482, + "learning_rate": 0.00023639158399803382, + "loss": 3.1593515872955322, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.031826410442590714, + "learning_rate": 0.0002361318150826927, + "loss": 3.1517653465270996, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.030141707509756088, + "learning_rate": 0.0002358721000418404, + "loss": 3.1291589736938477, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.03158571943640709, + "learning_rate": 0.00023561243908472472, + "loss": 3.178101062774658, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.03510458394885063, + "learning_rate": 0.00023535283242054954, + "loss": 3.138767719268799, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.03660859912633896, + "learning_rate": 0.00023509328025847565, + "loss": 3.1708762645721436, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.034004807472229004, + "learning_rate": 0.00023483378280761907, + "loss": 3.1404073238372803, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.03077496401965618, + "learning_rate": 0.0002345743402770527, + "loss": 3.159081220626831, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.030384650453925133, + "learning_rate": 0.0002343149528758042, + "loss": 3.1872847080230713, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.02957277186214924, + "learning_rate": 0.00023405562081285766, + "loss": 3.165973424911499, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.029529055580496788, + "learning_rate": 0.0002337963442971522, + "loss": 3.123840808868408, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.027776138857007027, + "learning_rate": 0.00023353712353758212, + "loss": 3.1705269813537598, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.028943249955773354, + "learning_rate": 0.00023327795874299704, + "loss": 3.1181654930114746, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.028178401291370392, + "learning_rate": 0.0002330188501222012, + "loss": 3.201678991317749, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.02865665778517723, + "learning_rate": 0.0002327597978839538, + "loss": 3.18479061126709, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.029173649847507477, + "learning_rate": 0.00023250080223696852, + "loss": 3.1844916343688965, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.0280170775949955, + "learning_rate": 0.00023224186338991352, + "loss": 3.187105417251587, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.029103878885507584, + "learning_rate": 0.0002319829815514112, + "loss": 3.183281660079956, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.029045486822724342, + "learning_rate": 0.00023172415693003789, + "loss": 3.1848196983337402, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.028364380821585655, + "learning_rate": 0.000231465389734324, + "loss": 3.197348117828369, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.029651399701833725, + "learning_rate": 0.00023120668017275346, + "loss": 3.1429920196533203, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.0281438659876585, + "learning_rate": 0.00023094802845376403, + "loss": 3.1916513442993164, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.029437677934765816, + "learning_rate": 0.00023068943478574662, + "loss": 3.172391176223755, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.03433044254779816, + "learning_rate": 0.00023043089937704541, + "loss": 3.2296864986419678, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.03259492665529251, + "learning_rate": 0.00023017242243595796, + "loss": 3.1962363719940186, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.0299630519002676, + "learning_rate": 0.00022991400417073427, + "loss": 3.169673442840576, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.028495769947767258, + "learning_rate": 0.00022965564478957739, + "loss": 3.1855409145355225, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.0290420800447464, + "learning_rate": 0.0002293973445006427, + "loss": 3.23667573928833, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.03340664133429527, + "learning_rate": 0.00022913910351203816, + "loss": 3.1873788833618164, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.03253626823425293, + "learning_rate": 0.00022888092203182377, + "loss": 3.204796552658081, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.028142794966697693, + "learning_rate": 0.00022862280026801182, + "loss": 3.193922281265259, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.028805622830986977, + "learning_rate": 0.00022836473842856623, + "loss": 3.1863901615142822, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.030033612623810768, + "learning_rate": 0.00022810673672140272, + "loss": 3.198482036590576, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.028154220432043076, + "learning_rate": 0.0002278487953543889, + "loss": 3.215069055557251, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.028098132461309433, + "learning_rate": 0.00022759091453534308, + "loss": 3.2197628021240234, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.028679659590125084, + "learning_rate": 0.00022733309447203553, + "loss": 3.2062196731567383, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.029104258865118027, + "learning_rate": 0.00022707533537218695, + "loss": 3.2182908058166504, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.029629778116941452, + "learning_rate": 0.00022681763744346938, + "loss": 3.2165639400482178, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.028304366394877434, + "learning_rate": 0.00022656000089350524, + "loss": 3.1783785820007324, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.026900488883256912, + "learning_rate": 0.00022630242592986766, + "loss": 3.1876602172851562, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.027047287672758102, + "learning_rate": 0.0002260449127600803, + "loss": 3.1879465579986572, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.027424108237028122, + "learning_rate": 0.0002257874615916166, + "loss": 3.2407851219177246, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.029319508001208305, + "learning_rate": 0.00022553007263190067, + "loss": 3.1980130672454834, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.027114372700452805, + "learning_rate": 0.00022527274608830574, + "loss": 3.211357593536377, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.02579840086400509, + "learning_rate": 0.0002250154821681555, + "loss": 3.2067861557006836, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.02767331339418888, + "learning_rate": 0.0002247582810787227, + "loss": 3.198587417602539, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.026645265519618988, + "learning_rate": 0.00022450114302722957, + "loss": 3.177441120147705, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.026181906461715698, + "learning_rate": 0.0002242440682208477, + "loss": 3.226815700531006, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.027218692004680634, + "learning_rate": 0.00022398705686669748, + "loss": 3.1731510162353516, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.027551919221878052, + "learning_rate": 0.0002237301091718485, + "loss": 3.2268621921539307, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.025782009586691856, + "learning_rate": 0.00022347322534331865, + "loss": 3.2052621841430664, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.024231620132923126, + "learning_rate": 0.00022321640558807488, + "loss": 3.201770544052124, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.02474607340991497, + "learning_rate": 0.00022295965011303188, + "loss": 3.1935834884643555, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.026299327611923218, + "learning_rate": 0.00022270295912505317, + "loss": 3.223275899887085, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.02439914271235466, + "learning_rate": 0.00022244633283095, + "loss": 3.2513954639434814, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.02650653012096882, + "learning_rate": 0.0002221897714374814, + "loss": 3.1914851665496826, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.024847298860549927, + "learning_rate": 0.00022193327515135445, + "loss": 3.2173104286193848, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.028475770726799965, + "learning_rate": 0.00022167684417922337, + "loss": 3.2184863090515137, + "step": 2160 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 9.208840083996672e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-2160/training_args.bin b/runs/l2r90-baseline/checkpoint-2160/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2160/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-2520/chat_template.jinja b/runs/l2r90-baseline/checkpoint-2520/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-2520/config.json b/runs/l2r90-baseline/checkpoint-2520/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-2520/generation_config.json b/runs/l2r90-baseline/checkpoint-2520/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-2520/model.safetensors b/runs/l2r90-baseline/checkpoint-2520/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9a395ee4cf04af48594b64d3a1216b531f6ad6ea --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a2d0d096c5824a8aa66c26ef12ef67dac462778badde06e4dbad522cf179a19 +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-2520/optimizer.pt b/runs/l2r90-baseline/checkpoint-2520/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..d749555ccbeb8066775e6261e820ab2f8596307b --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0b2388cb0aa77011ebb6d03d323ee12143d66821fc79bb665fa4113faf9f05c +size 1166825338 diff --git a/runs/l2r90-baseline/checkpoint-2520/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-2520/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..eaeeea885c2ef4b0cef75101a8e62f4ab4851bd2 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1e03ae1cb1232d1d92f456eeb7fed9b86dbea40002020bd21a8bff940afba04 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2520/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-2520/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..81543e2df97bda93c650637e0e76e6eba567cc34 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:204f99921454e515b3056a84d1cf89d690447610764df571b00287728068baff +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2520/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-2520/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..4ff4a5c1c578c912bb73e9e1039feddae85108d0 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9add74f0deb40ada69213c31c9eb2148f496716039a30fb746923520918c78da +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2520/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-2520/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..fc461818ec193e0f1784351202790a5fa22189ee --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff9312b009886e7daf125d940b7a052dd2e1a5c49cb82b56a0aaa5ca1c30fb9a +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2520/scheduler.pt b/runs/l2r90-baseline/checkpoint-2520/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..b53f13f877641b522f2c435934f587fdad6161cc --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f776ea81adcefd29eff4d7371e55373e8890bd750697e1c8b781d131e0a39a6 +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-2520/tokenizer.json b/runs/l2r90-baseline/checkpoint-2520/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-2520/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-2520/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-2520/trainer_state.json b/runs/l2r90-baseline/checkpoint-2520/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..3a1defa4787fc5278ed37bc19202dcd5bdc21204 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/trainer_state.json @@ -0,0 +1,17946 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 35.0, + "eval_steps": 500, + "global_step": 2520, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 5.549652099609375, + "eval_runtime": 57.761, + "eval_samples_per_second": 42.278, + "eval_steps_per_second": 1.333, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.03958690166473389, + "learning_rate": 0.0005919487226162231, + "loss": 5.459286689758301, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.03299890458583832, + "learning_rate": 0.0005918869546699877, + "loss": 5.45464563369751, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.028040863573551178, + "learning_rate": 0.0005918249539727509, + "loss": 5.444796562194824, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.028365720063447952, + "learning_rate": 0.0005917627205744655, + "loss": 5.447852611541748, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.02822224237024784, + "learning_rate": 0.0005917002545252719, + "loss": 5.453883171081543, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.036075398325920105, + "learning_rate": 0.0005916375558754977, + "loss": 5.462671279907227, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04683876410126686, + "learning_rate": 0.0005915746246756581, + "loss": 5.443917274475098, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.05457588657736778, + "learning_rate": 0.0005915114609764558, + "loss": 5.444374084472656, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.04250836744904518, + "learning_rate": 0.0005914480648287804, + "loss": 5.435029983520508, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.029477844014763832, + "learning_rate": 0.0005913844362837093, + "loss": 5.421862602233887, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.03673284500837326, + "learning_rate": 0.0005913205753925066, + "loss": 5.430391311645508, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.030489861965179443, + "learning_rate": 0.0005912564822066241, + "loss": 5.389314651489258, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.04037046805024147, + "learning_rate": 0.0005911921567777005, + "loss": 5.425060272216797, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.0445217490196228, + "learning_rate": 0.0005911275991575613, + "loss": 5.437565326690674, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.03885014355182648, + "learning_rate": 0.0005910628093982198, + "loss": 5.410930633544922, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.0334964245557785, + "learning_rate": 0.0005909977875518759, + "loss": 5.412092208862305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.03499286621809006, + "learning_rate": 0.0005909325336709164, + "loss": 5.382216453552246, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.04171644523739815, + "learning_rate": 0.0005908670478079152, + "loss": 5.387385368347168, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.057558078318834305, + "learning_rate": 0.000590801330015633, + "loss": 5.437925815582275, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.053846780210733414, + "learning_rate": 0.0005907353803470177, + "loss": 5.41644811630249, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.027259036898612976, + "learning_rate": 0.0005906691988552034, + "loss": 5.409185409545898, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.03394043818116188, + "learning_rate": 0.0005906027855935115, + "loss": 5.377973556518555, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.03731574863195419, + "learning_rate": 0.0005905361406154501, + "loss": 5.382505893707275, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.04705141484737396, + "learning_rate": 0.0005904692639747137, + "loss": 5.413661003112793, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.044811323285102844, + "learning_rate": 0.0005904021557251837, + "loss": 5.417835235595703, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.04000752419233322, + "learning_rate": 0.0005903348159209277, + "loss": 5.406160354614258, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.038364164531230927, + "learning_rate": 0.0005902672446162007, + "loss": 5.357950687408447, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.030500127002596855, + "learning_rate": 0.0005901994418654432, + "loss": 5.344951629638672, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.030644729733467102, + "learning_rate": 0.0005901314077232829, + "loss": 5.3599348068237305, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.037976790219545364, + "learning_rate": 0.0005900631422445335, + "loss": 5.354084491729736, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.04106011986732483, + "learning_rate": 0.0005899946454841955, + "loss": 5.363678455352783, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.04334832355380058, + "learning_rate": 0.0005899259174974553, + "loss": 5.350959300994873, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.049939028918743134, + "learning_rate": 0.000589856958339686, + "loss": 5.3317155838012695, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.05300058051943779, + "learning_rate": 0.0005897877680664465, + "loss": 5.353314399719238, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.05943184345960617, + "learning_rate": 0.0005897183467334822, + "loss": 5.3683929443359375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.05600789934396744, + "learning_rate": 0.0005896486943967249, + "loss": 5.345516204833984, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.05392718315124512, + "learning_rate": 0.0005895788111122919, + "loss": 5.291389465332031, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.037913475185632706, + "learning_rate": 0.0005895086969364871, + "loss": 5.353423118591309, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.03362123295664787, + "learning_rate": 0.0005894383519258001, + "loss": 5.379500389099121, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032558735460042953, + "learning_rate": 0.0005893677761369066, + "loss": 5.34043025970459, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.04074402526021004, + "learning_rate": 0.0005892969696266683, + "loss": 5.303869247436523, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.04270484298467636, + "learning_rate": 0.0005892259324521328, + "loss": 5.27297830581665, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.04604959115386009, + "learning_rate": 0.0005891546646705334, + "loss": 5.313268184661865, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.04462246224284172, + "learning_rate": 0.0005890831663392891, + "loss": 5.31167459487915, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.04966094344854355, + "learning_rate": 0.0005890114375160052, + "loss": 5.316695690155029, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.04662073776125908, + "learning_rate": 0.0005889394782584718, + "loss": 5.316898822784424, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.04380315542221069, + "learning_rate": 0.0005888672886246656, + "loss": 5.285087585449219, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03239579498767853, + "learning_rate": 0.0005887948686727483, + "loss": 5.304283142089844, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.029690183699131012, + "learning_rate": 0.0005887222184610675, + "loss": 5.305715084075928, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.03444967418909073, + "learning_rate": 0.0005886493380481559, + "loss": 5.2707672119140625, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.03302565962076187, + "learning_rate": 0.0005885762274927322, + "loss": 5.277251243591309, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.028477763757109642, + "learning_rate": 0.0005885028868537, + "loss": 5.295318603515625, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.022988801822066307, + "learning_rate": 0.0005884293161901487, + "loss": 5.249057292938232, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.026115216314792633, + "learning_rate": 0.000588355515561353, + "loss": 5.270636081695557, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.02576332725584507, + "learning_rate": 0.0005882814850267723, + "loss": 5.263096332550049, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.023978352546691895, + "learning_rate": 0.0005882072246460521, + "loss": 5.276605606079102, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.020695513114333153, + "learning_rate": 0.0005881327344790223, + "loss": 5.2141432762146, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01952219195663929, + "learning_rate": 0.0005880580145856986, + "loss": 5.214289665222168, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.022633206099271774, + "learning_rate": 0.0005879830650262813, + "loss": 5.236690998077393, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.023082347586750984, + "learning_rate": 0.0005879078858611557, + "loss": 5.2092156410217285, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.019620463252067566, + "learning_rate": 0.0005878324771508927, + "loss": 5.264451503753662, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.021504493430256844, + "learning_rate": 0.0005877568389562475, + "loss": 5.23444938659668, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.025672364979982376, + "learning_rate": 0.0005876809713381606, + "loss": 5.205289363861084, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.034436631947755814, + "learning_rate": 0.0005876048743577569, + "loss": 5.2047648429870605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.042389530688524246, + "learning_rate": 0.0005875285480763466, + "loss": 5.192008018493652, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.03310716897249222, + "learning_rate": 0.0005874519925554244, + "loss": 5.162734031677246, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.03029288910329342, + "learning_rate": 0.0005873752078566694, + "loss": 5.212151527404785, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04619310051202774, + "learning_rate": 0.0005872981940419459, + "loss": 5.215909004211426, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.05618087574839592, + "learning_rate": 0.0005872209511733027, + "loss": 5.210192680358887, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.03783707693219185, + "learning_rate": 0.0005871434793129726, + "loss": 5.201597690582275, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.03721104562282562, + "learning_rate": 0.0005870657785233735, + "loss": 5.188891410827637, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.033336155116558075, + "learning_rate": 0.0005869878488671075, + "loss": 5.1598615646362305, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.2592363357543945, + "eval_runtime": 57.3343, + "eval_samples_per_second": 42.592, + "eval_steps_per_second": 1.343, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03719741478562355, + "learning_rate": 0.0005869096904069611, + "loss": 5.1646623611450195, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.04888090118765831, + "learning_rate": 0.0005868313032059052, + "loss": 5.171182155609131, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.05073149502277374, + "learning_rate": 0.0005867526873270949, + "loss": 5.183949947357178, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.04187731072306633, + "learning_rate": 0.0005866738428338695, + "loss": 5.176133155822754, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.039445552974939346, + "learning_rate": 0.000586594769789753, + "loss": 5.148100852966309, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.035551708191633224, + "learning_rate": 0.0005865154682584524, + "loss": 5.151512145996094, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.031175289303064346, + "learning_rate": 0.0005864359383038602, + "loss": 5.102468967437744, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.03696789592504501, + "learning_rate": 0.000586356179990052, + "loss": 5.119902610778809, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.04836135357618332, + "learning_rate": 0.0005862761933812875, + "loss": 5.15183162689209, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.05206914246082306, + "learning_rate": 0.0005861959785420106, + "loss": 5.154870986938477, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.046368084847927094, + "learning_rate": 0.000586115535536849, + "loss": 5.130651473999023, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.07146432995796204, + "learning_rate": 0.000586034864430614, + "loss": 5.175804138183594, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.08809789270162582, + "learning_rate": 0.0005859539652883008, + "loss": 5.229716777801514, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.08116360008716583, + "learning_rate": 0.0005858728381750888, + "loss": 5.21910285949707, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.08282570540904999, + "learning_rate": 0.0005857914831563402, + "loss": 5.2336506843566895, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.07403547316789627, + "learning_rate": 0.0005857099002976016, + "loss": 5.251604080200195, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.07024004310369492, + "learning_rate": 0.0005856280896646026, + "loss": 5.267114162445068, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.056330062448978424, + "learning_rate": 0.0005855460513232567, + "loss": 5.201976299285889, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.051763955503702164, + "learning_rate": 0.0005854637853396606, + "loss": 5.1733198165893555, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.039900992065668106, + "learning_rate": 0.0005853812917800945, + "loss": 5.222941875457764, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.04187863692641258, + "learning_rate": 0.0005852985707110221, + "loss": 5.165033340454102, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.042645812034606934, + "learning_rate": 0.0005852156221990901, + "loss": 5.157554626464844, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.03856075555086136, + "learning_rate": 0.0005851324463111289, + "loss": 5.184629440307617, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.029387330636382103, + "learning_rate": 0.0005850490431141516, + "loss": 5.118739128112793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.027584781870245934, + "learning_rate": 0.0005849654126753545, + "loss": 5.100855350494385, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.023655924946069717, + "learning_rate": 0.0005848815550621175, + "loss": 5.154520034790039, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.024163050577044487, + "learning_rate": 0.0005847974703420028, + "loss": 5.105189323425293, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.020369920879602432, + "learning_rate": 0.0005847131585827564, + "loss": 5.113671779632568, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.021030737087130547, + "learning_rate": 0.0005846286198523061, + "loss": 5.0847649574279785, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.01729249767959118, + "learning_rate": 0.0005845438542187638, + "loss": 5.053225517272949, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.018704712390899658, + "learning_rate": 0.0005844588617504236, + "loss": 5.079106330871582, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.02049064077436924, + "learning_rate": 0.0005843736425157621, + "loss": 5.044185161590576, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.020736856386065483, + "learning_rate": 0.000584288196583439, + "loss": 5.041932106018066, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.016543276607990265, + "learning_rate": 0.0005842025240222966, + "loss": 5.032522201538086, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.018414106220006943, + "learning_rate": 0.0005841166249013598, + "loss": 5.037397384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.01611698977649212, + "learning_rate": 0.0005840304992898359, + "loss": 5.038179397583008, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.014310561120510101, + "learning_rate": 0.0005839441472571147, + "loss": 5.0146484375, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.015472214668989182, + "learning_rate": 0.0005838575688727685, + "loss": 5.060323715209961, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.016373401507735252, + "learning_rate": 0.0005837707642065518, + "loss": 5.032100677490234, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.018694665282964706, + "learning_rate": 0.000583683733328402, + "loss": 5.036993026733398, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.019977087154984474, + "learning_rate": 0.0005835964763084378, + "loss": 4.991037368774414, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.023741086944937706, + "learning_rate": 0.000583508993216961, + "loss": 5.043832778930664, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.025061562657356262, + "learning_rate": 0.0005834212841244548, + "loss": 5.016516208648682, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.020848384127020836, + "learning_rate": 0.0005833333491015851, + "loss": 4.986289978027344, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.015086354687809944, + "learning_rate": 0.0005832451882191995, + "loss": 4.9948625564575195, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.015096964314579964, + "learning_rate": 0.0005831568015483274, + "loss": 4.996342658996582, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016932200640439987, + "learning_rate": 0.0005830681891601807, + "loss": 4.998758316040039, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.019936930388212204, + "learning_rate": 0.0005829793511261525, + "loss": 4.959854602813721, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.027658682316541672, + "learning_rate": 0.0005828902875178183, + "loss": 4.968629837036133, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.03299275413155556, + "learning_rate": 0.0005828009984069347, + "loss": 4.994814872741699, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.033717650920152664, + "learning_rate": 0.0005827114838654406, + "loss": 4.976295471191406, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.02969924360513687, + "learning_rate": 0.0005826217439654562, + "loss": 4.969155788421631, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.02846827358007431, + "learning_rate": 0.0005825317787792831, + "loss": 4.989721298217773, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.03670600801706314, + "learning_rate": 0.0005824415883794049, + "loss": 4.988689422607422, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.045570675283670425, + "learning_rate": 0.0005823511728384863, + "loss": 4.974407196044922, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.04564111679792404, + "learning_rate": 0.0005822605322293733, + "loss": 4.966331481933594, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.04173210263252258, + "learning_rate": 0.0005821696666250937, + "loss": 4.925197601318359, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.04228242486715317, + "learning_rate": 0.0005820785760988559, + "loss": 4.965020656585693, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.05144782364368439, + "learning_rate": 0.0005819872607240503, + "loss": 5.0022149085998535, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.049400657415390015, + "learning_rate": 0.0005818957205742478, + "loss": 4.9635114669799805, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.043692708015441895, + "learning_rate": 0.0005818039557232005, + "loss": 4.977723121643066, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.04891742020845413, + "learning_rate": 0.0005817119662448421, + "loss": 4.996312141418457, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.035564981400966644, + "learning_rate": 0.0005816197522132866, + "loss": 4.940932273864746, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04679827764630318, + "learning_rate": 0.0005815273137028292, + "loss": 4.983131408691406, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.05489890277385712, + "learning_rate": 0.0005814346507879459, + "loss": 4.977827548980713, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04640809819102287, + "learning_rate": 0.0005813417635432937, + "loss": 4.954761505126953, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04439452290534973, + "learning_rate": 0.00058124865204371, + "loss": 4.914865970611572, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.04587096720933914, + "learning_rate": 0.0005811553163642131, + "loss": 4.949592590332031, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.044119734317064285, + "learning_rate": 0.0005810617565800019, + "loss": 4.956068992614746, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.043133001774549484, + "learning_rate": 0.0005809679727664559, + "loss": 4.993522644042969, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04315278306603432, + "learning_rate": 0.0005808739649991348, + "loss": 4.894291877746582, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04819472134113312, + "learning_rate": 0.0005807797333537792, + "loss": 4.958306789398193, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.023706436157227, + "eval_runtime": 57.4082, + "eval_samples_per_second": 42.537, + "eval_steps_per_second": 1.341, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.04403403028845787, + "learning_rate": 0.0005806852779063098, + "loss": 4.884971618652344, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.041993748396635056, + "learning_rate": 0.0005805905987328275, + "loss": 4.916353225708008, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.042302004992961884, + "learning_rate": 0.0005804956959096138, + "loss": 4.9012556076049805, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.044500596821308136, + "learning_rate": 0.0005804005695131299, + "loss": 4.915240287780762, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.03914736211299896, + "learning_rate": 0.0005803052196200177, + "loss": 4.885915756225586, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.0361703559756279, + "learning_rate": 0.0005802096463070988, + "loss": 4.829287528991699, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03432302549481392, + "learning_rate": 0.0005801138496513749, + "loss": 4.926834583282471, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.0317804180085659, + "learning_rate": 0.0005800178297300276, + "loss": 4.8707170486450195, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.035261351615190506, + "learning_rate": 0.0005799215866204183, + "loss": 4.897791862487793, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.03276067599654198, + "learning_rate": 0.0005798251204000886, + "loss": 4.864504814147949, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.034757066518068314, + "learning_rate": 0.0005797284311467594, + "loss": 4.890812873840332, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.033726081252098083, + "learning_rate": 0.0005796315189383316, + "loss": 4.831684112548828, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.03074726276099682, + "learning_rate": 0.0005795343838528855, + "loss": 4.8534111976623535, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02215893566608429, + "learning_rate": 0.0005794370259686811, + "loss": 4.8311238288879395, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.02132033370435238, + "learning_rate": 0.0005793394453641581, + "loss": 4.842617034912109, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020825816318392754, + "learning_rate": 0.0005792416421179351, + "loss": 4.84535026550293, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.019334357231855392, + "learning_rate": 0.0005791436163088106, + "loss": 4.839937686920166, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017372334375977516, + "learning_rate": 0.0005790453680157622, + "loss": 4.8115553855896, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.0190406683832407, + "learning_rate": 0.0005789468973179468, + "loss": 4.832542419433594, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.017372574657201767, + "learning_rate": 0.0005788482042947004, + "loss": 4.793632507324219, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.019171632826328278, + "learning_rate": 0.0005787492890255382, + "loss": 4.776987552642822, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.02109278179705143, + "learning_rate": 0.0005786501515901545, + "loss": 4.8216657638549805, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.026234494522213936, + "learning_rate": 0.0005785507920684225, + "loss": 4.8109049797058105, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.0357641875743866, + "learning_rate": 0.0005784512105403944, + "loss": 4.821109771728516, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.03627947345376015, + "learning_rate": 0.000578351407086301, + "loss": 4.810632228851318, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.032253146171569824, + "learning_rate": 0.0005782513817865527, + "loss": 4.7930521965026855, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.03402210772037506, + "learning_rate": 0.0005781511347217375, + "loss": 4.772500991821289, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.0321584977209568, + "learning_rate": 0.000578050665972623, + "loss": 4.781399726867676, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.030242707580327988, + "learning_rate": 0.0005779499756201549, + "loss": 4.802542686462402, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.02361336164176464, + "learning_rate": 0.0005778490637454576, + "loss": 4.771912574768066, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.028781302273273468, + "learning_rate": 0.000577747930429834, + "loss": 4.796924591064453, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.027042683213949203, + "learning_rate": 0.0005776465757547654, + "loss": 4.745218276977539, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.02715984545648098, + "learning_rate": 0.0005775449998019113, + "loss": 4.792944431304932, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.0314096100628376, + "learning_rate": 0.00057744320265311, + "loss": 4.7649993896484375, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.028749193996191025, + "learning_rate": 0.0005773411843903773, + "loss": 4.7980546951293945, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.02655309997498989, + "learning_rate": 0.0005772389450959075, + "loss": 4.746520519256592, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.027497397735714912, + "learning_rate": 0.000577136484852073, + "loss": 4.7490715980529785, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.0343393050134182, + "learning_rate": 0.000577033803741424, + "loss": 4.74436092376709, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.04444412514567375, + "learning_rate": 0.0005769309018466891, + "loss": 4.755973815917969, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.0437556728720665, + "learning_rate": 0.0005768277792507744, + "loss": 4.784368515014648, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0369921550154686, + "learning_rate": 0.0005767244360367638, + "loss": 4.753410339355469, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.05038141459226608, + "learning_rate": 0.0005766208722879192, + "loss": 4.755491256713867, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.04511422663927078, + "learning_rate": 0.00057651708808768, + "loss": 4.760293006896973, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.048762883991003036, + "learning_rate": 0.0005764130835196632, + "loss": 4.755958080291748, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.06749686598777771, + "learning_rate": 0.0005763088586676634, + "loss": 4.802557468414307, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.06451250612735748, + "learning_rate": 0.0005762044136156528, + "loss": 4.812925338745117, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.05436156690120697, + "learning_rate": 0.0005760997484477809, + "loss": 4.807103157043457, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.05233754590153694, + "learning_rate": 0.0005759948632483742, + "loss": 4.782449722290039, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.05519746616482735, + "learning_rate": 0.0005758897581019371, + "loss": 4.808449745178223, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.0578567236661911, + "learning_rate": 0.000575784433093151, + "loss": 4.7676005363464355, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.0588264986872673, + "learning_rate": 0.0005756788883068743, + "loss": 4.805377960205078, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04225790873169899, + "learning_rate": 0.0005755731238281423, + "loss": 4.748775005340576, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.0376368910074234, + "learning_rate": 0.0005754671397421678, + "loss": 4.759791374206543, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03830771893262863, + "learning_rate": 0.0005753609361343402, + "loss": 4.729974746704102, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.03684520348906517, + "learning_rate": 0.0005752545130902256, + "loss": 4.729854583740234, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02983408235013485, + "learning_rate": 0.0005751478706955674, + "loss": 4.78502082824707, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02854904904961586, + "learning_rate": 0.0005750410090362854, + "loss": 4.717472076416016, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.023819150403141975, + "learning_rate": 0.0005749339281984761, + "loss": 4.717124938964844, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.02477158047258854, + "learning_rate": 0.0005748266282684124, + "loss": 4.71781063079834, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.022002534940838814, + "learning_rate": 0.0005747191093325443, + "loss": 4.7480669021606445, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.02133483812212944, + "learning_rate": 0.0005746113714774976, + "loss": 4.693770408630371, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.022407138720154762, + "learning_rate": 0.0005745034147900747, + "loss": 4.720734119415283, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.020850926637649536, + "learning_rate": 0.0005743952393572544, + "loss": 4.745505332946777, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.020420582965016365, + "learning_rate": 0.0005742868452661918, + "loss": 4.688610076904297, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.019014107063412666, + "learning_rate": 0.0005741782326042181, + "loss": 4.673436164855957, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.017393039539456367, + "learning_rate": 0.0005740694014588403, + "loss": 4.6941118240356445, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.017877329140901566, + "learning_rate": 0.0005739603519177419, + "loss": 4.667474269866943, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.017749635502696037, + "learning_rate": 0.0005738510840687821, + "loss": 4.6901092529296875, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.018746376037597656, + "learning_rate": 0.000573741597999996, + "loss": 4.680166721343994, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016260217875242233, + "learning_rate": 0.0005736318937995947, + "loss": 4.6799821853637695, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015099814161658287, + "learning_rate": 0.0005735219715559646, + "loss": 4.680666923522949, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.016539735719561577, + "learning_rate": 0.0005734118313576683, + "loss": 4.639509201049805, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 4.768754482269287, + "eval_runtime": 58.3352, + "eval_samples_per_second": 41.862, + "eval_steps_per_second": 1.32, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.019733518362045288, + "learning_rate": 0.0005733014732934436, + "loss": 4.635312080383301, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.022571036592125893, + "learning_rate": 0.0005731908974522042, + "loss": 4.6041765213012695, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.022238552570343018, + "learning_rate": 0.0005730801039230389, + "loss": 4.648976802825928, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.02115180902183056, + "learning_rate": 0.0005729690927952123, + "loss": 4.623968124389648, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.021789928898215294, + "learning_rate": 0.0005728578641581637, + "loss": 4.620028018951416, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.02593887411057949, + "learning_rate": 0.0005727464181015081, + "loss": 4.591062545776367, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.0347842313349247, + "learning_rate": 0.0005726347547150357, + "loss": 4.632149696350098, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04119519516825676, + "learning_rate": 0.0005725228740887117, + "loss": 4.6254191398620605, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.04895750433206558, + "learning_rate": 0.0005724107763126761, + "loss": 4.621372222900391, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.05752161890268326, + "learning_rate": 0.0005722984614772442, + "loss": 4.61965274810791, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05115870013833046, + "learning_rate": 0.000572185929672906, + "loss": 4.63119649887085, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.05721089243888855, + "learning_rate": 0.0005720731809903263, + "loss": 4.652647018432617, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.0646839588880539, + "learning_rate": 0.000571960215520345, + "loss": 4.642184257507324, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.0682854950428009, + "learning_rate": 0.0005718470333539757, + "loss": 4.676589012145996, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.057402849197387695, + "learning_rate": 0.0005717336345824077, + "loss": 4.6585283279418945, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.05885373800992966, + "learning_rate": 0.0005716200192970043, + "loss": 4.671728134155273, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.05265705659985542, + "learning_rate": 0.0005715061875893032, + "loss": 4.641249656677246, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.05417842045426369, + "learning_rate": 0.0005713921395510166, + "loss": 4.661381721496582, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.06602369248867035, + "learning_rate": 0.0005712778752740307, + "loss": 4.684137344360352, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.06004182621836662, + "learning_rate": 0.0005711633948504066, + "loss": 4.625332832336426, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.04801511764526367, + "learning_rate": 0.0005710486983723787, + "loss": 4.664791107177734, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.04380105435848236, + "learning_rate": 0.0005709337859323561, + "loss": 4.664966583251953, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.03858727216720581, + "learning_rate": 0.0005708186576229218, + "loss": 4.661867618560791, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03240576758980751, + "learning_rate": 0.0005707033135368323, + "loss": 4.5855326652526855, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.028561925515532494, + "learning_rate": 0.0005705877537670184, + "loss": 4.637331008911133, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.026389246806502342, + "learning_rate": 0.0005704719784065846, + "loss": 4.598666667938232, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.02484819106757641, + "learning_rate": 0.0005703559875488088, + "loss": 4.601729393005371, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.02522977627813816, + "learning_rate": 0.0005702397812871429, + "loss": 4.609833717346191, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02589508146047592, + "learning_rate": 0.0005701233597152121, + "loss": 4.574592590332031, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.023374728858470917, + "learning_rate": 0.0005700067229268154, + "loss": 4.61362361907959, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.02465851791203022, + "learning_rate": 0.0005698898710159245, + "loss": 4.639899253845215, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.02258715033531189, + "learning_rate": 0.0005697728040766852, + "loss": 4.5656538009643555, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.018009789288043976, + "learning_rate": 0.0005696555222034162, + "loss": 4.61449670791626, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02121281623840332, + "learning_rate": 0.0005695380254906094, + "loss": 4.553830623626709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.018078701570630074, + "learning_rate": 0.0005694203140329296, + "loss": 4.534579277038574, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.0169303547590971, + "learning_rate": 0.000569302387925215, + "loss": 4.533995628356934, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.016318142414093018, + "learning_rate": 0.0005691842472624764, + "loss": 4.544438362121582, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.016664575785398483, + "learning_rate": 0.0005690658921398977, + "loss": 4.5913286209106445, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.01641504280269146, + "learning_rate": 0.0005689473226528354, + "loss": 4.520743370056152, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.017012735828757286, + "learning_rate": 0.0005688285388968187, + "loss": 4.517453193664551, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01672312617301941, + "learning_rate": 0.0005687095409675499, + "loss": 4.5235915184021, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.016893865540623665, + "learning_rate": 0.000568590328960903, + "loss": 4.543200492858887, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.017530135810375214, + "learning_rate": 0.0005684709029729253, + "loss": 4.54912805557251, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.018698599189519882, + "learning_rate": 0.0005683512630998361, + "loss": 4.522650718688965, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01984071172773838, + "learning_rate": 0.000568231409438027, + "loss": 4.546173095703125, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.023218289017677307, + "learning_rate": 0.0005681113420840619, + "loss": 4.5305094718933105, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.028968270868062973, + "learning_rate": 0.000567991061134677, + "loss": 4.521440505981445, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.032370273023843765, + "learning_rate": 0.0005678705666867805, + "loss": 4.54168701171875, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.033967748284339905, + "learning_rate": 0.0005677498588374524, + "loss": 4.538717269897461, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.028757929801940918, + "learning_rate": 0.0005676289376839452, + "loss": 4.517228603363037, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.029757825657725334, + "learning_rate": 0.0005675078033236827, + "loss": 4.510310173034668, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.03581003099679947, + "learning_rate": 0.0005673864558542605, + "loss": 4.506901264190674, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.03483985736966133, + "learning_rate": 0.0005672648953734462, + "loss": 4.549284934997559, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.031016027554869652, + "learning_rate": 0.000567143121979179, + "loss": 4.530084609985352, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.029373526573181152, + "learning_rate": 0.0005670211357695693, + "loss": 4.492218971252441, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.031661372631788254, + "learning_rate": 0.0005668989368428994, + "loss": 4.502824783325195, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.027251387014985085, + "learning_rate": 0.0005667765252976227, + "loss": 4.49193811416626, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02263176068663597, + "learning_rate": 0.0005666539012323639, + "loss": 4.512019157409668, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.02355228364467621, + "learning_rate": 0.0005665310647459189, + "loss": 4.508917331695557, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.023297371342778206, + "learning_rate": 0.0005664080159372551, + "loss": 4.497771263122559, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.019364114850759506, + "learning_rate": 0.0005662847549055107, + "loss": 4.51470947265625, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.023899901658296585, + "learning_rate": 0.0005661612817499947, + "loss": 4.48771858215332, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.02330983616411686, + "learning_rate": 0.0005660375965701872, + "loss": 4.478496551513672, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.027252234518527985, + "learning_rate": 0.0005659136994657392, + "loss": 4.497672080993652, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.030886363238096237, + "learning_rate": 0.0005657895905364723, + "loss": 4.49789571762085, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02515559457242489, + "learning_rate": 0.0005656652698823788, + "loss": 4.490771293640137, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020107125863432884, + "learning_rate": 0.0005655407376036217, + "loss": 4.481879234313965, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.022076716646552086, + "learning_rate": 0.0005654159938005345, + "loss": 4.494697570800781, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.025525815784931183, + "learning_rate": 0.0005652910385736207, + "loss": 4.517453193664551, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.03005894646048546, + "learning_rate": 0.0005651658720235548, + "loss": 4.453815460205078, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.030178502202033997, + "learning_rate": 0.0005650404942511812, + "loss": 4.493922233581543, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.029249394312500954, + "learning_rate": 0.0005649149053575141, + "loss": 4.453763008117676, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 4.607339382171631, + "eval_runtime": 57.9216, + "eval_samples_per_second": 42.16, + "eval_steps_per_second": 1.329, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.03362950682640076, + "learning_rate": 0.0005647891054437386, + "loss": 4.430700302124023, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.04223684221506119, + "learning_rate": 0.0005646630946112096, + "loss": 4.486915588378906, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.05192171037197113, + "learning_rate": 0.0005645368729614512, + "loss": 4.467541694641113, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.042525868862867355, + "learning_rate": 0.0005644104405961582, + "loss": 4.45252799987793, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.0406370609998703, + "learning_rate": 0.0005642837976171949, + "loss": 4.414548873901367, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.044476233422756195, + "learning_rate": 0.0005641569441265952, + "loss": 4.453492164611816, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.04687326401472092, + "learning_rate": 0.0005640298802265626, + "loss": 4.477725028991699, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.04011611267924309, + "learning_rate": 0.0005639026060194704, + "loss": 4.423356533050537, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.04971884936094284, + "learning_rate": 0.0005637751216078609, + "loss": 4.4853515625, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.05863234028220177, + "learning_rate": 0.0005636474270944461, + "loss": 4.466001033782959, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.057710036635398865, + "learning_rate": 0.0005635195225821072, + "loss": 4.466723442077637, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.05155074968934059, + "learning_rate": 0.0005633914081738945, + "loss": 4.4823198318481445, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.04535634070634842, + "learning_rate": 0.0005632630839730275, + "loss": 4.471219062805176, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.045147854834795, + "learning_rate": 0.0005631345500828946, + "loss": 4.476888179779053, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.03730619326233864, + "learning_rate": 0.0005630058066070533, + "loss": 4.428932189941406, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.034599728882312775, + "learning_rate": 0.0005628768536492299, + "loss": 4.47627592086792, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.032904356718063354, + "learning_rate": 0.0005627476913133193, + "loss": 4.422671794891357, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.03356337174773216, + "learning_rate": 0.0005626183197033856, + "loss": 4.456406593322754, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.035042595118284225, + "learning_rate": 0.0005624887389236609, + "loss": 4.371820449829102, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.03269032761454582, + "learning_rate": 0.0005623589490785462, + "loss": 4.402605056762695, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.028261031955480576, + "learning_rate": 0.0005622289502726108, + "loss": 4.433136940002441, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03301700949668884, + "learning_rate": 0.0005620987426105925, + "loss": 4.388359546661377, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.031664177775382996, + "learning_rate": 0.000561968326197397, + "loss": 4.41859245300293, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.026281699538230896, + "learning_rate": 0.0005618377011380988, + "loss": 4.406128883361816, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.023097911849617958, + "learning_rate": 0.0005617068675379398, + "loss": 4.389717102050781, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.022349286824464798, + "learning_rate": 0.0005615758255023306, + "loss": 4.408173561096191, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.021958226338028908, + "learning_rate": 0.0005614445751368491, + "loss": 4.434830188751221, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024551713839173317, + "learning_rate": 0.0005613131165472415, + "loss": 4.386944770812988, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.028048967942595482, + "learning_rate": 0.0005611814498394216, + "loss": 4.402279376983643, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02733907848596573, + "learning_rate": 0.0005610495751194708, + "loss": 4.403069496154785, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.020858559757471085, + "learning_rate": 0.0005609174924936384, + "loss": 4.406920909881592, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.021491149440407753, + "learning_rate": 0.0005607852020683409, + "loss": 4.385799407958984, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.021495329216122627, + "learning_rate": 0.0005606527039501621, + "loss": 4.4252495765686035, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.019708355888724327, + "learning_rate": 0.0005605199982458535, + "loss": 4.388864994049072, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.01886274851858616, + "learning_rate": 0.0005603870850623338, + "loss": 4.335962295532227, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.01955341547727585, + "learning_rate": 0.0005602539645066884, + "loss": 4.373560905456543, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.018962012603878975, + "learning_rate": 0.0005601206366861706, + "loss": 4.380022048950195, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.016663335263729095, + "learning_rate": 0.0005599871017081999, + "loss": 4.421034336090088, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01676873490214348, + "learning_rate": 0.0005598533596803631, + "loss": 4.382775783538818, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.014379126951098442, + "learning_rate": 0.0005597194107104137, + "loss": 4.41373872756958, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.015897583216428757, + "learning_rate": 0.000559585254906272, + "loss": 4.377284049987793, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.016513120383024216, + "learning_rate": 0.000559450892376025, + "loss": 4.41200065612793, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.016625918447971344, + "learning_rate": 0.000559316323227926, + "loss": 4.381187438964844, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.016028188169002533, + "learning_rate": 0.000559181547570395, + "loss": 4.353909015655518, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01682914048433304, + "learning_rate": 0.0005590465655120183, + "loss": 4.382326126098633, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.017078889533877373, + "learning_rate": 0.0005589113771615486, + "loss": 4.334949493408203, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.018182938918471336, + "learning_rate": 0.0005587759826279046, + "loss": 4.348130226135254, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01940886303782463, + "learning_rate": 0.0005586403820201713, + "loss": 4.339404106140137, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.020644692704081535, + "learning_rate": 0.0005585045754475995, + "loss": 4.358404159545898, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02433430217206478, + "learning_rate": 0.0005583685630196064, + "loss": 4.3660502433776855, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.027824537828564644, + "learning_rate": 0.0005582323448457744, + "loss": 4.34050178527832, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.02973582223057747, + "learning_rate": 0.0005580959210358523, + "loss": 4.348322868347168, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.030396128073334694, + "learning_rate": 0.0005579592916997542, + "loss": 4.356224060058594, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.029760530218482018, + "learning_rate": 0.0005578224569475599, + "loss": 4.357470989227295, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.02855013869702816, + "learning_rate": 0.0005576854168895147, + "loss": 4.335353374481201, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03376635164022446, + "learning_rate": 0.0005575481716360292, + "loss": 4.331794738769531, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.034431129693984985, + "learning_rate": 0.0005574107212976796, + "loss": 4.336880207061768, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.03524135425686836, + "learning_rate": 0.000557273065985207, + "loss": 4.31537389755249, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.03152058273553848, + "learning_rate": 0.0005571352058095179, + "loss": 4.37662410736084, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02878640592098236, + "learning_rate": 0.0005569971408816838, + "loss": 4.292388916015625, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.029663654044270515, + "learning_rate": 0.0005568588713129409, + "loss": 4.344346523284912, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.027895759791135788, + "learning_rate": 0.0005567203972146906, + "loss": 4.365072727203369, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.026465769857168198, + "learning_rate": 0.000556581718698499, + "loss": 4.334959506988525, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.03191852569580078, + "learning_rate": 0.000556442835876097, + "loss": 4.361069679260254, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.03535444661974907, + "learning_rate": 0.0005563037488593799, + "loss": 4.349765777587891, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03451581299304962, + "learning_rate": 0.0005561644577604074, + "loss": 4.343142509460449, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029810845851898193, + "learning_rate": 0.000556024962691404, + "loss": 4.307485580444336, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.027523217722773552, + "learning_rate": 0.0005558852637647584, + "loss": 4.366406440734863, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03043208085000515, + "learning_rate": 0.0005557453610930235, + "loss": 4.356929779052734, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.03058922104537487, + "learning_rate": 0.0005556052547889161, + "loss": 4.348942279815674, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.032052166759967804, + "learning_rate": 0.0005554649449653177, + "loss": 4.3496809005737305, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.03170694038271904, + "learning_rate": 0.0005553244317352731, + "loss": 4.315959930419922, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 4.4757890701293945, + "eval_runtime": 57.7451, + "eval_samples_per_second": 42.289, + "eval_steps_per_second": 1.333, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.029025450348854065, + "learning_rate": 0.0005551837152119915, + "loss": 4.288647174835205, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.029403962194919586, + "learning_rate": 0.0005550427955088455, + "loss": 4.288555145263672, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.03265019878745079, + "learning_rate": 0.0005549016727393715, + "loss": 4.292625904083252, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.0387202650308609, + "learning_rate": 0.0005547603470172697, + "loss": 4.238926887512207, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.04554206505417824, + "learning_rate": 0.0005546188184564036, + "loss": 4.29966402053833, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.036231476813554764, + "learning_rate": 0.0005544770871708004, + "loss": 4.304398536682129, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.038703009486198425, + "learning_rate": 0.0005543351532746501, + "loss": 4.2945356369018555, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03224619850516319, + "learning_rate": 0.0005541930168823067, + "loss": 4.261351585388184, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.029722856357693672, + "learning_rate": 0.0005540506781082864, + "loss": 4.303143501281738, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.02764045260846615, + "learning_rate": 0.0005539081370672695, + "loss": 4.243679523468018, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.0312200877815485, + "learning_rate": 0.0005537653938740984, + "loss": 4.279904365539551, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.027349065989255905, + "learning_rate": 0.0005536224486437788, + "loss": 4.270228385925293, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.030385233461856842, + "learning_rate": 0.0005534793014914793, + "loss": 4.301590442657471, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.03292274847626686, + "learning_rate": 0.0005533359525325307, + "loss": 4.301828384399414, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.03265977278351784, + "learning_rate": 0.0005531924018824266, + "loss": 4.286200523376465, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.03177836537361145, + "learning_rate": 0.0005530486496568236, + "loss": 4.296319007873535, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.028879934921860695, + "learning_rate": 0.0005529046959715397, + "loss": 4.261706829071045, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.02855812944471836, + "learning_rate": 0.0005527605409425562, + "loss": 4.273903846740723, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03037901036441326, + "learning_rate": 0.0005526161846860161, + "loss": 4.248080253601074, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.029279105365276337, + "learning_rate": 0.0005524716273182245, + "loss": 4.281311988830566, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.02592580020427704, + "learning_rate": 0.0005523268689556489, + "loss": 4.204848289489746, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.021899603307247162, + "learning_rate": 0.0005521819097149183, + "loss": 4.232453346252441, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.019704442471265793, + "learning_rate": 0.0005520367497128238, + "loss": 4.253462314605713, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.019550519064068794, + "learning_rate": 0.0005518913890663182, + "loss": 4.307737350463867, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.020320113748311996, + "learning_rate": 0.0005517458278925161, + "loss": 4.257411956787109, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02118559181690216, + "learning_rate": 0.0005516000663086932, + "loss": 4.262089729309082, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.02050616219639778, + "learning_rate": 0.0005514541044322872, + "loss": 4.2490644454956055, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019793255254626274, + "learning_rate": 0.0005513079423808967, + "loss": 4.267400741577148, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.020760254934430122, + "learning_rate": 0.0005511615802722823, + "loss": 4.2532429695129395, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.019524799659848213, + "learning_rate": 0.0005510150182243647, + "loss": 4.267422199249268, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.020166637375950813, + "learning_rate": 0.0005508682563552269, + "loss": 4.214296817779541, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.01964394748210907, + "learning_rate": 0.0005507212947831119, + "loss": 4.252087116241455, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01690433733165264, + "learning_rate": 0.0005505741336264239, + "loss": 4.232336044311523, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.017682762816548347, + "learning_rate": 0.0005504267730037283, + "loss": 4.213962554931641, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01688133366405964, + "learning_rate": 0.0005502792130337508, + "loss": 4.199461460113525, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.019166061654686928, + "learning_rate": 0.0005501314538353776, + "loss": 4.228936672210693, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.024065518751740456, + "learning_rate": 0.000549983495527656, + "loss": 4.214227676391602, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02518288791179657, + "learning_rate": 0.000549835338229793, + "loss": 4.256412029266357, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.026093967258930206, + "learning_rate": 0.0005496869820611563, + "loss": 4.216564178466797, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.023775635287165642, + "learning_rate": 0.0005495384271412739, + "loss": 4.226430892944336, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.018810907378792763, + "learning_rate": 0.0005493896735898338, + "loss": 4.190013885498047, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.018974872305989265, + "learning_rate": 0.000549240721526684, + "loss": 4.246231555938721, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.02069833129644394, + "learning_rate": 0.0005490915710718325, + "loss": 4.198457717895508, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.02097306028008461, + "learning_rate": 0.0005489422223454471, + "loss": 4.212507247924805, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022165542468428612, + "learning_rate": 0.0005487926754678556, + "loss": 4.2043843269348145, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.019068919122219086, + "learning_rate": 0.0005486429305595448, + "loss": 4.197460651397705, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01701812446117401, + "learning_rate": 0.0005484929877411618, + "loss": 4.220056533813477, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01629726029932499, + "learning_rate": 0.0005483428471335127, + "loss": 4.214868545532227, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018661806359887123, + "learning_rate": 0.000548192508857563, + "loss": 4.227047920227051, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.02005891129374504, + "learning_rate": 0.0005480419730344377, + "loss": 4.197265625, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.02591128647327423, + "learning_rate": 0.0005478912397854208, + "loss": 4.211041450500488, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.032659754157066345, + "learning_rate": 0.0005477403092319552, + "loss": 4.230195045471191, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.04057808965444565, + "learning_rate": 0.0005475891814956429, + "loss": 4.223843097686768, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.04103454574942589, + "learning_rate": 0.0005474378566982449, + "loss": 4.197145462036133, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.03810729831457138, + "learning_rate": 0.0005472863349616809, + "loss": 4.237957000732422, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.028838301077485085, + "learning_rate": 0.0005471346164080292, + "loss": 4.206722259521484, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02436191774904728, + "learning_rate": 0.0005469827011595266, + "loss": 4.1983137130737305, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.022839387878775597, + "learning_rate": 0.0005468305893385686, + "loss": 4.20530366897583, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.02375260926783085, + "learning_rate": 0.0005466782810677088, + "loss": 4.213202476501465, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.02177741937339306, + "learning_rate": 0.0005465257764696593, + "loss": 4.200539588928223, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.02351468615233898, + "learning_rate": 0.0005463730756672903, + "loss": 4.198091506958008, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.023375794291496277, + "learning_rate": 0.0005462201787836301, + "loss": 4.193244934082031, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.02479805052280426, + "learning_rate": 0.0005460670859418651, + "loss": 4.2250871658325195, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.026894841343164444, + "learning_rate": 0.000545913797265339, + "loss": 4.241791248321533, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.030582403764128685, + "learning_rate": 0.0005457603128775541, + "loss": 4.203666687011719, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.032222259789705276, + "learning_rate": 0.0005456066329021699, + "loss": 4.209629535675049, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.03181060031056404, + "learning_rate": 0.0005454527574630037, + "loss": 4.200459957122803, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02687576413154602, + "learning_rate": 0.0005452986866840298, + "loss": 4.211819648742676, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.025988668203353882, + "learning_rate": 0.0005451444206893806, + "loss": 4.198040962219238, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.02937682718038559, + "learning_rate": 0.0005449899596033453, + "loss": 4.222935676574707, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.03043951839208603, + "learning_rate": 0.0005448353035503704, + "loss": 4.2031755447387695, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.030401621013879776, + "learning_rate": 0.0005446804526550596, + "loss": 4.198380470275879, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 4.368444919586182, + "eval_runtime": 58.3009, + "eval_samples_per_second": 41.886, + "eval_steps_per_second": 1.321, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.03461585193872452, + "learning_rate": 0.0005445254070421732, + "loss": 4.147369384765625, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.039084456861019135, + "learning_rate": 0.0005443701668366288, + "loss": 4.124129295349121, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.04021848738193512, + "learning_rate": 0.0005442147321635007, + "loss": 4.127362251281738, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.041000768542289734, + "learning_rate": 0.00054405910314802, + "loss": 4.1432061195373535, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.04008613899350166, + "learning_rate": 0.0005439032799155737, + "loss": 4.169931888580322, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.039394307881593704, + "learning_rate": 0.000543747262591706, + "loss": 4.155813217163086, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.04374868795275688, + "learning_rate": 0.0005435910513021172, + "loss": 4.159891128540039, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0443582646548748, + "learning_rate": 0.0005434346461726641, + "loss": 4.157362937927246, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.04601646214723587, + "learning_rate": 0.0005432780473293593, + "loss": 4.152097225189209, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.04327242821455002, + "learning_rate": 0.0005431212548983718, + "loss": 4.170074462890625, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.0413462370634079, + "learning_rate": 0.0005429642690060262, + "loss": 4.156710147857666, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.036855220794677734, + "learning_rate": 0.0005428070897788034, + "loss": 4.138053894042969, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.03781217709183693, + "learning_rate": 0.0005426497173433398, + "loss": 4.172306060791016, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.03886301815509796, + "learning_rate": 0.0005424921518264275, + "loss": 4.135544776916504, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.03730994835495949, + "learning_rate": 0.0005423343933550143, + "loss": 4.171654224395752, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.03271950036287308, + "learning_rate": 0.0005421764420562032, + "loss": 4.169906139373779, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.03292673081159592, + "learning_rate": 0.0005420182980572527, + "loss": 4.163660049438477, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.029366906732320786, + "learning_rate": 0.0005418599614855768, + "loss": 4.161055088043213, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.023805437609553337, + "learning_rate": 0.0005417014324687442, + "loss": 4.172390937805176, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.023178137838840485, + "learning_rate": 0.000541542711134479, + "loss": 4.146788597106934, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.02403450198471546, + "learning_rate": 0.0005413837976106599, + "loss": 4.124364852905273, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.025009717792272568, + "learning_rate": 0.0005412246920253208, + "loss": 4.12275505065918, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.023431289941072464, + "learning_rate": 0.0005410653945066503, + "loss": 4.134739398956299, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.023803815245628357, + "learning_rate": 0.0005409059051829914, + "loss": 4.130463123321533, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.021219994872808456, + "learning_rate": 0.0005407462241828417, + "loss": 4.124577522277832, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.01899666152894497, + "learning_rate": 0.0005405863516348535, + "loss": 4.167086124420166, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.01889493502676487, + "learning_rate": 0.000540426287667833, + "loss": 4.152744293212891, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.019546914845705032, + "learning_rate": 0.000540266032410741, + "loss": 4.122655868530273, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02085448056459427, + "learning_rate": 0.0005401055859926923, + "loss": 4.141949653625488, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.019959403201937675, + "learning_rate": 0.0005399449485429555, + "loss": 4.131907939910889, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.01885392889380455, + "learning_rate": 0.0005397841201909535, + "loss": 4.118953227996826, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.018216131255030632, + "learning_rate": 0.0005396231010662627, + "loss": 4.096459865570068, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.019385207444429398, + "learning_rate": 0.0005394618912986132, + "loss": 4.111050128936768, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.018676603212952614, + "learning_rate": 0.0005393004910178889, + "loss": 4.156215667724609, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.01876339502632618, + "learning_rate": 0.0005391389003541271, + "loss": 4.1107282638549805, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.01831623539328575, + "learning_rate": 0.0005389771194375185, + "loss": 4.154116630554199, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.018896661698818207, + "learning_rate": 0.0005388151483984071, + "loss": 4.093320846557617, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.019582601264119148, + "learning_rate": 0.0005386529873672898, + "loss": 4.106500625610352, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.020313601940870285, + "learning_rate": 0.0005384906364748172, + "loss": 4.071961879730225, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.020385954529047012, + "learning_rate": 0.000538328095851792, + "loss": 4.129006862640381, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02231123298406601, + "learning_rate": 0.0005381653656291707, + "loss": 4.067756652832031, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.022957179695367813, + "learning_rate": 0.0005380024459380619, + "loss": 4.088230133056641, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021438462659716606, + "learning_rate": 0.000537839336909727, + "loss": 4.128036022186279, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01893841102719307, + "learning_rate": 0.00053767603867558, + "loss": 4.155407905578613, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.018539676442742348, + "learning_rate": 0.0005375125513671874, + "loss": 4.101880073547363, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.01881185919046402, + "learning_rate": 0.0005373488751162681, + "loss": 4.118906021118164, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.021657396107912064, + "learning_rate": 0.0005371850100546929, + "loss": 4.091935157775879, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.023293187841773033, + "learning_rate": 0.000537020956314485, + "loss": 4.105803489685059, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.021306317299604416, + "learning_rate": 0.0005368567140278196, + "loss": 4.0610809326171875, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.022078553214669228, + "learning_rate": 0.0005366922833270235, + "loss": 4.09867525100708, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.02266664244234562, + "learning_rate": 0.0005365276643445757, + "loss": 4.1027326583862305, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.02091757394373417, + "learning_rate": 0.0005363628572131069, + "loss": 4.080266952514648, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.020374659448862076, + "learning_rate": 0.0005361978620653991, + "loss": 4.088479042053223, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.02294750325381756, + "learning_rate": 0.0005360326790343858, + "loss": 4.084364891052246, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.02337353676557541, + "learning_rate": 0.000535867308253152, + "loss": 4.125696182250977, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.022643696516752243, + "learning_rate": 0.0005357017498549341, + "loss": 4.097160339355469, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.021080605685710907, + "learning_rate": 0.0005355360039731196, + "loss": 4.102363109588623, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.02269427664577961, + "learning_rate": 0.0005353700707412466, + "loss": 4.089844703674316, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.02195413038134575, + "learning_rate": 0.0005352039502930048, + "loss": 4.132293701171875, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.01805616170167923, + "learning_rate": 0.0005350376427622343, + "loss": 4.0842814445495605, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.018048783764243126, + "learning_rate": 0.0005348711482829261, + "loss": 4.109591007232666, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.0199508648365736, + "learning_rate": 0.000534704466989222, + "loss": 4.0908074378967285, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.020186785608530045, + "learning_rate": 0.0005345375990154138, + "loss": 4.077959060668945, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.022526193410158157, + "learning_rate": 0.0005343705444959441, + "loss": 4.109979629516602, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.025224441662430763, + "learning_rate": 0.000534203303565406, + "loss": 4.047533988952637, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.027292724698781967, + "learning_rate": 0.000534035876358542, + "loss": 4.09566593170166, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.025669090449810028, + "learning_rate": 0.0005338682630102454, + "loss": 4.0606160163879395, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02791150100529194, + "learning_rate": 0.0005337004636555595, + "loss": 4.097081184387207, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.03294133022427559, + "learning_rate": 0.0005335324784296768, + "loss": 4.121442794799805, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.027322066947817802, + "learning_rate": 0.0005333643074679403, + "loss": 4.1188740730285645, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.024866372346878052, + "learning_rate": 0.000533195950905842, + "loss": 4.082042694091797, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.02901027724146843, + "learning_rate": 0.0005330274088790239, + "loss": 4.110199451446533, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 4.283970355987549, + "eval_runtime": 58.4003, + "eval_samples_per_second": 41.815, + "eval_steps_per_second": 1.318, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.028958842158317566, + "learning_rate": 0.0005328586815232773, + "loss": 4.040733337402344, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.03491180017590523, + "learning_rate": 0.0005326897689745428, + "loss": 4.0473432540893555, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.04025818780064583, + "learning_rate": 0.0005325206713689101, + "loss": 4.021976470947266, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03998899459838867, + "learning_rate": 0.0005323513888426181, + "loss": 4.04901123046875, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.042165111750364304, + "learning_rate": 0.0005321819215320546, + "loss": 4.050230979919434, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.03511597216129303, + "learning_rate": 0.0005320122695737565, + "loss": 4.0989990234375, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.04064979776740074, + "learning_rate": 0.0005318424331044094, + "loss": 4.051448822021484, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.047536592930555344, + "learning_rate": 0.0005316724122608473, + "loss": 4.064278602600098, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.05592669919133186, + "learning_rate": 0.0005315022071800527, + "loss": 4.045516014099121, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.04262633994221687, + "learning_rate": 0.0005313318179991574, + "loss": 4.060666084289551, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03997781500220299, + "learning_rate": 0.0005311612448554402, + "loss": 4.065315246582031, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.038152433931827545, + "learning_rate": 0.0005309904878863293, + "loss": 4.069721221923828, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.0367623008787632, + "learning_rate": 0.0005308195472294001, + "loss": 4.080986976623535, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.0358598455786705, + "learning_rate": 0.0005306484230223769, + "loss": 4.065290927886963, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.039188165217638016, + "learning_rate": 0.000530477115403131, + "loss": 4.024333477020264, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.036839909851551056, + "learning_rate": 0.0005303056245096817, + "loss": 4.067532539367676, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.03456985577940941, + "learning_rate": 0.0005301339504801967, + "loss": 4.048443794250488, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.039487943053245544, + "learning_rate": 0.0005299620934529902, + "loss": 4.0585737228393555, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.03415827825665474, + "learning_rate": 0.0005297900535665244, + "loss": 4.06104850769043, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.03209620714187622, + "learning_rate": 0.0005296178309594087, + "loss": 4.009165287017822, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.028947360813617706, + "learning_rate": 0.0005294454257704, + "loss": 4.067384243011475, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026836825534701347, + "learning_rate": 0.0005292728381384018, + "loss": 4.046419620513916, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024095216765999794, + "learning_rate": 0.0005291000682024647, + "loss": 4.019008636474609, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.020965851843357086, + "learning_rate": 0.0005289271161017867, + "loss": 4.020613670349121, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.02250855043530464, + "learning_rate": 0.0005287539819757119, + "loss": 4.031634330749512, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.021385207772254944, + "learning_rate": 0.0005285806659637315, + "loss": 4.086246013641357, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.021104589104652405, + "learning_rate": 0.000528407168205483, + "loss": 4.033970355987549, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.019584037363529205, + "learning_rate": 0.0005282334888407503, + "loss": 4.025118827819824, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.02107258513569832, + "learning_rate": 0.000528059628009464, + "loss": 4.020823955535889, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.020438389852643013, + "learning_rate": 0.0005278855858517003, + "loss": 4.016323566436768, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.01881488971412182, + "learning_rate": 0.0005277113625076822, + "loss": 3.9991841316223145, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.01763618178665638, + "learning_rate": 0.000527536958117778, + "loss": 4.0062150955200195, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.017308862879872322, + "learning_rate": 0.0005273623728225022, + "loss": 3.946723461151123, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01706845872104168, + "learning_rate": 0.0005271876067625153, + "loss": 4.056562900543213, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.018664294853806496, + "learning_rate": 0.0005270126600786228, + "loss": 4.024652481079102, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.018523404374718666, + "learning_rate": 0.0005268375329117762, + "loss": 3.962064743041992, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.017408935353159904, + "learning_rate": 0.0005266622254030724, + "loss": 4.008697509765625, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.01588570512831211, + "learning_rate": 0.0005264867376937534, + "loss": 4.011664390563965, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.015084310434758663, + "learning_rate": 0.0005263110699252065, + "loss": 3.996224880218506, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01637054979801178, + "learning_rate": 0.000526135222238964, + "loss": 4.015972137451172, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.01673802360892296, + "learning_rate": 0.0005259591947767032, + "loss": 3.975414276123047, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.016544019803404808, + "learning_rate": 0.0005257829876802464, + "loss": 4.012955188751221, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.017455460503697395, + "learning_rate": 0.0005256066010915603, + "loss": 4.010180473327637, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.017625460401177406, + "learning_rate": 0.0005254300351527566, + "loss": 4.001679420471191, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.01732795126736164, + "learning_rate": 0.000525253290006091, + "loss": 4.006107330322266, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.016654646024107933, + "learning_rate": 0.0005250763657939641, + "loss": 4.045180797576904, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.015085658989846706, + "learning_rate": 0.0005248992626589203, + "loss": 3.994246482849121, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.015112568624317646, + "learning_rate": 0.0005247219807436489, + "loss": 4.004963397979736, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.014882124960422516, + "learning_rate": 0.000524544520190982, + "loss": 4.029565811157227, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.016012471169233322, + "learning_rate": 0.000524366881143897, + "loss": 4.018208026885986, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.015761924907565117, + "learning_rate": 0.0005241890637455141, + "loss": 3.9911205768585205, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.017770355567336082, + "learning_rate": 0.0005240110681390978, + "loss": 4.028733253479004, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01959751546382904, + "learning_rate": 0.0005238328944680558, + "loss": 4.003322124481201, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.018713824450969696, + "learning_rate": 0.0005236545428759394, + "loss": 3.991718292236328, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.01992208883166313, + "learning_rate": 0.0005234760135064434, + "loss": 4.002507209777832, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.021671291440725327, + "learning_rate": 0.0005232973065034055, + "loss": 4.021344184875488, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024282798171043396, + "learning_rate": 0.0005231184220108069, + "loss": 3.9899158477783203, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.023294344544410706, + "learning_rate": 0.0005229393601727713, + "loss": 4.046627044677734, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.02029707282781601, + "learning_rate": 0.000522760121133566, + "loss": 4.022454261779785, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.02151239849627018, + "learning_rate": 0.0005225807050376004, + "loss": 4.005476951599121, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.02320881187915802, + "learning_rate": 0.0005224011120294269, + "loss": 4.012726783752441, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.026484191417694092, + "learning_rate": 0.0005222213422537401, + "loss": 3.9904391765594482, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.026279661804437637, + "learning_rate": 0.0005220413958553777, + "loss": 3.9738848209381104, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.024299729615449905, + "learning_rate": 0.0005218612729793188, + "loss": 4.003316402435303, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.02337595447897911, + "learning_rate": 0.0005216809737706856, + "loss": 3.954984188079834, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.02416428178548813, + "learning_rate": 0.0005215004983747415, + "loss": 4.0317063331604, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.022488107904791832, + "learning_rate": 0.0005213198469368925, + "loss": 3.9736647605895996, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.019700225442647934, + "learning_rate": 0.0005211390196026861, + "loss": 3.979867458343506, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.020382333546876907, + "learning_rate": 0.0005209580165178117, + "loss": 3.995251178741455, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.020149298012256622, + "learning_rate": 0.0005207768378281003, + "loss": 3.9987268447875977, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.019790509715676308, + "learning_rate": 0.0005205954836795241, + "loss": 3.9838991165161133, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.022096451371908188, + "learning_rate": 0.000520413954218197, + "loss": 3.995800256729126, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.214347839355469, + "eval_runtime": 58.8872, + "eval_samples_per_second": 41.469, + "eval_steps_per_second": 1.308, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02203402668237686, + "learning_rate": 0.0005202322495903739, + "loss": 3.915332078933716, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.02336675301194191, + "learning_rate": 0.0005200503699424509, + "loss": 3.914801597595215, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.027339637279510498, + "learning_rate": 0.0005198683154209654, + "loss": 3.9277780055999756, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.026872653514146805, + "learning_rate": 0.0005196860861725952, + "loss": 3.9160094261169434, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.026558255776762962, + "learning_rate": 0.0005195036823441592, + "loss": 3.9084300994873047, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.02644367143511772, + "learning_rate": 0.0005193211040826169, + "loss": 3.9328718185424805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.024677276611328125, + "learning_rate": 0.0005191383515350681, + "loss": 3.88993501663208, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.023194244131445885, + "learning_rate": 0.0005189554248487536, + "loss": 3.923567056655884, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.024653147906064987, + "learning_rate": 0.0005187723241710539, + "loss": 3.8740999698638916, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.026041202247142792, + "learning_rate": 0.0005185890496494899, + "loss": 3.9542148113250732, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.029586758464574814, + "learning_rate": 0.0005184056014317225, + "loss": 3.9241347312927246, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.0325787179172039, + "learning_rate": 0.0005182219796655528, + "loss": 3.9656593799591064, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.03183380514383316, + "learning_rate": 0.0005180381844989214, + "loss": 3.950648307800293, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.03434471786022186, + "learning_rate": 0.0005178542160799088, + "loss": 3.9590439796447754, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.030099432915449142, + "learning_rate": 0.0005176700745567353, + "loss": 3.9518113136291504, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.025950564071536064, + "learning_rate": 0.0005174857600777599, + "loss": 3.8908939361572266, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.029959935694932938, + "learning_rate": 0.0005173012727914816, + "loss": 3.945037364959717, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.029551763087511063, + "learning_rate": 0.0005171166128465389, + "loss": 3.923959732055664, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.03538523241877556, + "learning_rate": 0.0005169317803917083, + "loss": 3.9395527839660645, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.037924155592918396, + "learning_rate": 0.0005167467755759065, + "loss": 3.945115089416504, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.036773014813661575, + "learning_rate": 0.0005165615985481881, + "loss": 3.9484646320343018, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.04125838354229927, + "learning_rate": 0.0005163762494577474, + "loss": 4.000699043273926, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.044125135987997055, + "learning_rate": 0.0005161907284539165, + "loss": 3.9828028678894043, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.040853384882211685, + "learning_rate": 0.0005160050356861663, + "loss": 3.957505702972412, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.03501008450984955, + "learning_rate": 0.0005158191713041063, + "loss": 3.9360849857330322, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.04082556813955307, + "learning_rate": 0.0005156331354574838, + "loss": 3.95941424369812, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.045432962477207184, + "learning_rate": 0.0005154469282961847, + "loss": 3.9666922092437744, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.04048244655132294, + "learning_rate": 0.0005152605499702328, + "loss": 3.984555244445801, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.031016314402222633, + "learning_rate": 0.0005150740006297896, + "loss": 3.945204257965088, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.03210771828889847, + "learning_rate": 0.0005148872804251548, + "loss": 3.976186752319336, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.03130023553967476, + "learning_rate": 0.000514700389506765, + "loss": 3.915727138519287, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.031193193048238754, + "learning_rate": 0.0005145133280251954, + "loss": 3.961866855621338, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.027518954128026962, + "learning_rate": 0.0005143260961311578, + "loss": 3.9742956161499023, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.02751155197620392, + "learning_rate": 0.0005141386939755014, + "loss": 3.949662208557129, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.026033353060483932, + "learning_rate": 0.0005139511217092129, + "loss": 3.9327926635742188, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.02579621411859989, + "learning_rate": 0.000513763379483416, + "loss": 3.9334418773651123, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.023009559139609337, + "learning_rate": 0.0005135754674493709, + "loss": 3.969707489013672, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.02196447364985943, + "learning_rate": 0.0005133873857584753, + "loss": 3.9084434509277344, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.02206484042108059, + "learning_rate": 0.0005131991345622629, + "loss": 3.922117233276367, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.023360205814242363, + "learning_rate": 0.0005130107140124046, + "loss": 3.965158224105835, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02162669226527214, + "learning_rate": 0.0005128221242607072, + "loss": 3.9519729614257812, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.019719846546649933, + "learning_rate": 0.0005126333654591141, + "loss": 3.9195961952209473, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.018988260999321938, + "learning_rate": 0.0005124444377597049, + "loss": 3.937926769256592, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.019278401508927345, + "learning_rate": 0.0005122553413146951, + "loss": 3.9126808643341064, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019101005047559738, + "learning_rate": 0.0005120660762764366, + "loss": 3.8993749618530273, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.019156700000166893, + "learning_rate": 0.0005118766427974166, + "loss": 3.9531068801879883, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019944407045841217, + "learning_rate": 0.0005116870410302582, + "loss": 3.9645180702209473, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.01826174184679985, + "learning_rate": 0.0005114972711277202, + "loss": 3.948477268218994, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.0191368218511343, + "learning_rate": 0.0005113073332426968, + "loss": 3.944648027420044, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.017768343910574913, + "learning_rate": 0.0005111172275282174, + "loss": 3.9584474563598633, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.01955590397119522, + "learning_rate": 0.0005109269541374469, + "loss": 3.8936800956726074, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.02100536972284317, + "learning_rate": 0.000510736513223685, + "loss": 3.93919038772583, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.020873935893177986, + "learning_rate": 0.0005105459049403665, + "loss": 3.929244041442871, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.021876046434044838, + "learning_rate": 0.0005103551294410614, + "loss": 3.9264731407165527, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.021726684644818306, + "learning_rate": 0.0005101641868794736, + "loss": 3.923135280609131, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.019308188930153847, + "learning_rate": 0.0005099730774094423, + "loss": 3.9062135219573975, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01917017623782158, + "learning_rate": 0.0005097818011849408, + "loss": 3.898810386657715, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017933931201696396, + "learning_rate": 0.0005095903583600772, + "loss": 3.8899734020233154, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.015131795778870583, + "learning_rate": 0.0005093987490890933, + "loss": 3.9452900886535645, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.017190825194120407, + "learning_rate": 0.0005092069735263651, + "loss": 3.9124746322631836, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.01618487387895584, + "learning_rate": 0.000509015031826403, + "loss": 3.91654109954834, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.014920198358595371, + "learning_rate": 0.0005088229241438509, + "loss": 3.9144535064697266, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01615205965936184, + "learning_rate": 0.0005086306506334864, + "loss": 3.917003631591797, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.015370192006230354, + "learning_rate": 0.0005084382114502208, + "loss": 3.8910603523254395, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.015290996059775352, + "learning_rate": 0.0005082456067490988, + "loss": 3.9122695922851562, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.015913106501102448, + "learning_rate": 0.0005080528366852987, + "loss": 3.9304556846618652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.01639600843191147, + "learning_rate": 0.0005078599014141318, + "loss": 3.8974502086639404, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.017010601237416267, + "learning_rate": 0.0005076668010910425, + "loss": 3.9723148345947266, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.017778996378183365, + "learning_rate": 0.0005074735358716083, + "loss": 3.8905391693115234, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.020029611885547638, + "learning_rate": 0.0005072801059115394, + "loss": 3.934976100921631, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.01931857503950596, + "learning_rate": 0.000507086511366679, + "loss": 3.9152774810791016, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019750485196709633, + "learning_rate": 0.0005068927523930027, + "loss": 3.877650499343872, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.161914825439453, + "eval_runtime": 58.5217, + "eval_samples_per_second": 41.728, + "eval_steps_per_second": 1.316, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02073049545288086, + "learning_rate": 0.0005066988291466184, + "loss": 3.8331246376037598, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02272389829158783, + "learning_rate": 0.0005065047417837669, + "loss": 3.8847155570983887, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02139078453183174, + "learning_rate": 0.0005063104904608206, + "loss": 3.8276169300079346, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02043229155242443, + "learning_rate": 0.0005061160753342843, + "loss": 3.8804001808166504, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.02345564216375351, + "learning_rate": 0.0005059214965607948, + "loss": 3.851609468460083, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.023448185995221138, + "learning_rate": 0.0005057267542971207, + "loss": 3.8446507453918457, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.024826066568493843, + "learning_rate": 0.0005055318487001624, + "loss": 3.86712384223938, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.02491980977356434, + "learning_rate": 0.0005053367799269518, + "loss": 3.8374528884887695, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.023002834990620613, + "learning_rate": 0.0005051415481346522, + "loss": 3.783214569091797, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.02123952843248844, + "learning_rate": 0.0005049461534805585, + "loss": 3.8463428020477295, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.01981225423514843, + "learning_rate": 0.0005047505961220964, + "loss": 3.8307082653045654, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01915367692708969, + "learning_rate": 0.0005045548762168231, + "loss": 3.8190226554870605, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.01890859752893448, + "learning_rate": 0.0005043589939224266, + "loss": 3.8428220748901367, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.019150376319885254, + "learning_rate": 0.0005041629493967257, + "loss": 3.830247402191162, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.020099453628063202, + "learning_rate": 0.0005039667427976697, + "loss": 3.838806629180908, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.021136781200766563, + "learning_rate": 0.0005037703742833391, + "loss": 3.8344273567199707, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.022857150062918663, + "learning_rate": 0.0005035738440119442, + "loss": 3.8317880630493164, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.024891739711165428, + "learning_rate": 0.0005033771521418261, + "loss": 3.837855577468872, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.02595464698970318, + "learning_rate": 0.0005031802988314557, + "loss": 3.846050262451172, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.02722409926354885, + "learning_rate": 0.0005029832842394342, + "loss": 3.8164305686950684, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.024486860260367393, + "learning_rate": 0.0005027861085244927, + "loss": 3.8099875450134277, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.021551169455051422, + "learning_rate": 0.0005025887718454922, + "loss": 3.825925827026367, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.02121506817638874, + "learning_rate": 0.0005023912743614232, + "loss": 3.870727777481079, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.01948230341076851, + "learning_rate": 0.0005021936162314058, + "loss": 3.8035314083099365, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.01896044984459877, + "learning_rate": 0.0005019957976146898, + "loss": 3.84665584564209, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.020620400086045265, + "learning_rate": 0.0005017978186706538, + "loss": 3.8338310718536377, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.020855262875556946, + "learning_rate": 0.0005015996795588062, + "loss": 3.8781003952026367, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.02153870463371277, + "learning_rate": 0.0005014013804387838, + "loss": 3.8783490657806396, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.023751024156808853, + "learning_rate": 0.0005012029214703529, + "loss": 3.846198558807373, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.023987066000699997, + "learning_rate": 0.000501004302813408, + "loss": 3.834296226501465, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.023626182228326797, + "learning_rate": 0.0005008055246279727, + "loss": 3.8591227531433105, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.025707192718982697, + "learning_rate": 0.000500606587074199, + "loss": 3.8766489028930664, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.027723029255867004, + "learning_rate": 0.0005004074903123674, + "loss": 3.88264799118042, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.0278895553201437, + "learning_rate": 0.0005002082345028864, + "loss": 3.8932995796203613, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.025622662156820297, + "learning_rate": 0.0005000088198062929, + "loss": 3.8734612464904785, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.02639615908265114, + "learning_rate": 0.0004998092463832516, + "loss": 3.786813735961914, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.026765599846839905, + "learning_rate": 0.0004996095143945554, + "loss": 3.853966474533081, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02823319099843502, + "learning_rate": 0.0004994096240011246, + "loss": 3.8321340084075928, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.023394955322146416, + "learning_rate": 0.0004992095753640072, + "loss": 3.8527331352233887, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.023874662816524506, + "learning_rate": 0.0004990093686443791, + "loss": 3.8822879791259766, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.02301311306655407, + "learning_rate": 0.000498809004003543, + "loss": 3.846186876296997, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.022630592808127403, + "learning_rate": 0.0004986084816029291, + "loss": 3.867992639541626, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.02424517087638378, + "learning_rate": 0.0004984078016040946, + "loss": 3.867293357849121, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.024116624146699905, + "learning_rate": 0.000498206964168724, + "loss": 3.846282958984375, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.02187768742442131, + "learning_rate": 0.000498005969458628, + "loss": 3.8429484367370605, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.01879809983074665, + "learning_rate": 0.0004978048176357447, + "loss": 3.863215446472168, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.020426234230399132, + "learning_rate": 0.0004976035088621385, + "loss": 3.825230121612549, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01992812193930149, + "learning_rate": 0.0004974020433000003, + "loss": 3.851665496826172, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.020693328231573105, + "learning_rate": 0.000497200421111647, + "loss": 3.906999111175537, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.021137919276952744, + "learning_rate": 0.0004969986424595221, + "loss": 3.8369035720825195, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.023285407572984695, + "learning_rate": 0.0004967967075061949, + "loss": 3.8658390045166016, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.0219135619699955, + "learning_rate": 0.0004965946164143609, + "loss": 3.873072862625122, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.020101983100175858, + "learning_rate": 0.0004963923693468413, + "loss": 3.847805976867676, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.01875176839530468, + "learning_rate": 0.0004961899664665827, + "loss": 3.7976622581481934, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01913360133767128, + "learning_rate": 0.0004959874079366576, + "loss": 3.853851318359375, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.016863619908690453, + "learning_rate": 0.0004957846939202637, + "loss": 3.877378225326538, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.017802366986870766, + "learning_rate": 0.000495581824580724, + "loss": 3.861356019973755, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.01981814205646515, + "learning_rate": 0.0004953788000814866, + "loss": 3.88613224029541, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.017508897930383682, + "learning_rate": 0.000495175620586125, + "loss": 3.8202030658721924, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.019240548834204674, + "learning_rate": 0.0004949722862583368, + "loss": 3.853827714920044, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.022305598482489586, + "learning_rate": 0.000494768797261945, + "loss": 3.885148763656616, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.02531025931239128, + "learning_rate": 0.000494565153760897, + "loss": 3.8412041664123535, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.024474849924445152, + "learning_rate": 0.0004943613559192648, + "loss": 3.865644931793213, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.020430924370884895, + "learning_rate": 0.0004941574039012444, + "loss": 3.8649442195892334, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.01906431093811989, + "learning_rate": 0.0004939532978711566, + "loss": 3.8462204933166504, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.021118057891726494, + "learning_rate": 0.0004937490379934456, + "loss": 3.8652522563934326, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.020337693393230438, + "learning_rate": 0.00049354462443268, + "loss": 3.8506221771240234, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.019090797752141953, + "learning_rate": 0.0004933400573535521, + "loss": 3.881237506866455, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.020973797887563705, + "learning_rate": 0.000493135336920878, + "loss": 3.821274757385254, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.021984362974762917, + "learning_rate": 0.0004929304632995973, + "loss": 3.8793835639953613, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.0215684212744236, + "learning_rate": 0.0004927254366547727, + "loss": 3.8259663581848145, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.021673865616321564, + "learning_rate": 0.0004925202571515908, + "loss": 3.8505711555480957, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.123871326446533, + "eval_runtime": 58.8357, + "eval_samples_per_second": 41.505, + "eval_steps_per_second": 1.309, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.0203359667211771, + "learning_rate": 0.0004923149249553608, + "loss": 3.785353660583496, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.023197868838906288, + "learning_rate": 0.0004921094402315152, + "loss": 3.7479515075683594, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.02314298041164875, + "learning_rate": 0.0004919038031456093, + "loss": 3.7697315216064453, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.0256908368319273, + "learning_rate": 0.0004916980138633213, + "loss": 3.79056978225708, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.026193121448159218, + "learning_rate": 0.0004914920725504518, + "loss": 3.7722532749176025, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.027120303362607956, + "learning_rate": 0.000491285979372924, + "loss": 3.7991981506347656, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.0266975536942482, + "learning_rate": 0.0004910797344967834, + "loss": 3.778301477432251, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.02806396596133709, + "learning_rate": 0.000490873338088198, + "loss": 3.7578704357147217, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.02893839031457901, + "learning_rate": 0.0004906667903134573, + "loss": 3.7401275634765625, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.031170370057225227, + "learning_rate": 0.0004904600913389735, + "loss": 3.780686616897583, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.03364266827702522, + "learning_rate": 0.0004902532413312799, + "loss": 3.794442653656006, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.030121006071567535, + "learning_rate": 0.0004900462404570321, + "loss": 3.779569149017334, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.03296204283833504, + "learning_rate": 0.0004898390888830067, + "loss": 3.7960336208343506, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.03381749242544174, + "learning_rate": 0.0004896317867761022, + "loss": 3.7959280014038086, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.035383883863687515, + "learning_rate": 0.000489424334303338, + "loss": 3.8008625507354736, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.03759761154651642, + "learning_rate": 0.0004892167316318548, + "loss": 3.7997028827667236, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.030651796609163284, + "learning_rate": 0.0004890089789289147, + "loss": 3.81191349029541, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.0262477844953537, + "learning_rate": 0.0004888010763618997, + "loss": 3.8046622276306152, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.030398324131965637, + "learning_rate": 0.0004885930240983136, + "loss": 3.7903637886047363, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.030515259131789207, + "learning_rate": 0.0004883848223057801, + "loss": 3.8111231327056885, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.031686048954725266, + "learning_rate": 0.00048817647115204386, + "loss": 3.781445026397705, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.025968654081225395, + "learning_rate": 0.00048796797080496936, + "loss": 3.788196563720703, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.025340190157294273, + "learning_rate": 0.00048775932143254176, + "loss": 3.763331413269043, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.024738596752285957, + "learning_rate": 0.0004875505232028661, + "loss": 3.8005969524383545, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.021990075707435608, + "learning_rate": 0.0004873415762841672, + "loss": 3.795844078063965, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.02191784791648388, + "learning_rate": 0.00048713248084478986, + "loss": 3.8062384128570557, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.023678259924054146, + "learning_rate": 0.0004869232370531986, + "loss": 3.7760679721832275, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.024137988686561584, + "learning_rate": 0.0004867138450779773, + "loss": 3.7917492389678955, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.02673223614692688, + "learning_rate": 0.0004865043050878293, + "loss": 3.8006649017333984, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.02395169623196125, + "learning_rate": 0.0004862946172515771, + "loss": 3.7902212142944336, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.02410844899713993, + "learning_rate": 0.0004860847817381625, + "loss": 3.8073291778564453, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.028109831735491753, + "learning_rate": 0.0004858747987166463, + "loss": 3.80116605758667, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.025538964197039604, + "learning_rate": 0.00048566466835620774, + "loss": 3.791151523590088, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.02393815852701664, + "learning_rate": 0.0004854543908261453, + "loss": 3.7766106128692627, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.024659741669893265, + "learning_rate": 0.0004852439662958756, + "loss": 3.789135217666626, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.02668776921927929, + "learning_rate": 0.000485033394934934, + "loss": 3.797905445098877, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.0258641317486763, + "learning_rate": 0.000484822676912974, + "loss": 3.783323287963867, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.024846022948622704, + "learning_rate": 0.0004846118123997674, + "loss": 3.7976717948913574, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.024981874972581863, + "learning_rate": 0.00048440080156520387, + "loss": 3.818732738494873, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.02556881681084633, + "learning_rate": 0.0004841896445792909, + "loss": 3.7821896076202393, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.02558571845293045, + "learning_rate": 0.000483978341612154, + "loss": 3.8143680095672607, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.02361566200852394, + "learning_rate": 0.0004837668928340362, + "loss": 3.7710189819335938, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022570716217160225, + "learning_rate": 0.0004835552984152978, + "loss": 3.775564670562744, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.022478004917502403, + "learning_rate": 0.00048334355852641664, + "loss": 3.7920689582824707, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021241728216409683, + "learning_rate": 0.0004831316733379879, + "loss": 3.7857837677001953, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.018921511247754097, + "learning_rate": 0.0004829196430207235, + "loss": 3.7659268379211426, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01926037296652794, + "learning_rate": 0.00048270746774545237, + "loss": 3.7527129650115967, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.018743667751550674, + "learning_rate": 0.0004824951476831204, + "loss": 3.767991065979004, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.017273949459195137, + "learning_rate": 0.00048228268300479015, + "loss": 3.7834174633026123, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.017054244875907898, + "learning_rate": 0.0004820700738816403, + "loss": 3.7775752544403076, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017190931364893913, + "learning_rate": 0.0004818573204849664, + "loss": 3.7709648609161377, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01665034517645836, + "learning_rate": 0.00048164442298617987, + "loss": 3.7575321197509766, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.016932843253016472, + "learning_rate": 0.00048143138155680844, + "loss": 3.7922799587249756, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01702750287950039, + "learning_rate": 0.0004812181963684958, + "loss": 3.7729780673980713, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01766599342226982, + "learning_rate": 0.00048100486759300134, + "loss": 3.7876572608947754, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01785697042942047, + "learning_rate": 0.00048079139540220024, + "loss": 3.8131699562072754, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01830688677728176, + "learning_rate": 0.0004805777799680831, + "loss": 3.790778636932373, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017348438501358032, + "learning_rate": 0.0004803640214627561, + "loss": 3.7984330654144287, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017999377101659775, + "learning_rate": 0.0004801501200584406, + "loss": 3.7756147384643555, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.01801210641860962, + "learning_rate": 0.000479936075927473, + "loss": 3.789759874343872, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.018445398658514023, + "learning_rate": 0.000479721889242305, + "loss": 3.7683000564575195, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.017544370144605637, + "learning_rate": 0.0004795075601755027, + "loss": 3.7776038646698, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.019834233447909355, + "learning_rate": 0.00047929308889974737, + "loss": 3.7559571266174316, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01791626401245594, + "learning_rate": 0.0004790784755878346, + "loss": 3.7746729850769043, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.017638562247157097, + "learning_rate": 0.00047886372041267434, + "loss": 3.7933082580566406, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.018974050879478455, + "learning_rate": 0.0004786488235472909, + "loss": 3.795713424682617, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.018694303929805756, + "learning_rate": 0.0004784337851648231, + "loss": 3.7675955295562744, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.01788092777132988, + "learning_rate": 0.0004782186054385233, + "loss": 3.786083698272705, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.01902497000992298, + "learning_rate": 0.0004780032845417579, + "loss": 3.7976233959198, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.020558880642056465, + "learning_rate": 0.000477787822648007, + "loss": 3.7500152587890625, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.020093899220228195, + "learning_rate": 0.00047757221993086443, + "loss": 3.703448534011841, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.019833799451589584, + "learning_rate": 0.00047735647656403754, + "loss": 3.756059408187866, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.093987941741943, + "eval_runtime": 57.6894, + "eval_samples_per_second": 42.33, + "eval_steps_per_second": 1.335, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.022167526185512543, + "learning_rate": 0.0004771405927213467, + "loss": 3.696444511413574, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.027874166145920753, + "learning_rate": 0.0004769245685767255, + "loss": 3.713630199432373, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.030872784554958344, + "learning_rate": 0.00047670840430422113, + "loss": 3.678295612335205, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.03246859833598137, + "learning_rate": 0.0004764921000779928, + "loss": 3.7186388969421387, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.03667069226503372, + "learning_rate": 0.0004762756560723131, + "loss": 3.7457995414733887, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.036286015063524246, + "learning_rate": 0.0004760590724615671, + "loss": 3.6881775856018066, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.03413614630699158, + "learning_rate": 0.0004758423494202522, + "loss": 3.6999263763427734, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.03384552150964737, + "learning_rate": 0.0004756254871229782, + "loss": 3.6959803104400635, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.03227181360125542, + "learning_rate": 0.00047540848574446724, + "loss": 3.7182722091674805, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.036366235464811325, + "learning_rate": 0.0004751913454595535, + "loss": 3.7116851806640625, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.03573326766490936, + "learning_rate": 0.0004749740664431828, + "loss": 3.741865873336792, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.03482096269726753, + "learning_rate": 0.000474756648870413, + "loss": 3.7195520401000977, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.03232625126838684, + "learning_rate": 0.0004745390929164134, + "loss": 3.7443251609802246, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.03421664610505104, + "learning_rate": 0.0004743213987564651, + "loss": 3.6867361068725586, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.03585013002157211, + "learning_rate": 0.0004741035665659603, + "loss": 3.7492403984069824, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.029985329136252403, + "learning_rate": 0.00047388559652040236, + "loss": 3.705209493637085, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.02729589305818081, + "learning_rate": 0.0004736674887954059, + "loss": 3.711106538772583, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.029642216861248016, + "learning_rate": 0.0004734492435666963, + "loss": 3.7274727821350098, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.02816903032362461, + "learning_rate": 0.0004732308610101099, + "loss": 3.7181742191314697, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.027143625542521477, + "learning_rate": 0.00047301234130159346, + "loss": 3.719235897064209, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.025802744552493095, + "learning_rate": 0.00047279368461720445, + "loss": 3.771820545196533, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.026424532756209373, + "learning_rate": 0.0004725748911331106, + "loss": 3.7516651153564453, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.029677648097276688, + "learning_rate": 0.0004723559610255898, + "loss": 3.7058472633361816, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.031548649072647095, + "learning_rate": 0.0004721368944710302, + "loss": 3.756992816925049, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.027667395770549774, + "learning_rate": 0.00047191769164592974, + "loss": 3.742711067199707, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.027639545500278473, + "learning_rate": 0.000471698352726896, + "loss": 3.728883743286133, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.02950332500040531, + "learning_rate": 0.0004714788778906467, + "loss": 3.7160212993621826, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.026302607730031013, + "learning_rate": 0.0004712592673140084, + "loss": 3.735713481903076, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.022483112290501595, + "learning_rate": 0.00047103952117391764, + "loss": 3.702686309814453, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.021729422733187675, + "learning_rate": 0.00047081963964741986, + "loss": 3.7195816040039062, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.021159254014492035, + "learning_rate": 0.0004705996229116696, + "loss": 3.684321880340576, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.02233424223959446, + "learning_rate": 0.0004703794711439304, + "loss": 3.7233057022094727, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.02142048068344593, + "learning_rate": 0.0004701591845215744, + "loss": 3.7083680629730225, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.022331461310386658, + "learning_rate": 0.00046993876322208274, + "loss": 3.7408461570739746, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.020629866048693657, + "learning_rate": 0.0004697182074230448, + "loss": 3.759129047393799, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.02174685336649418, + "learning_rate": 0.0004694975173021586, + "loss": 3.72351336479187, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.021538110449910164, + "learning_rate": 0.0004692766930372299, + "loss": 3.716416597366333, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.020800625905394554, + "learning_rate": 0.00046905573480617285, + "loss": 3.726954936981201, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.02039588987827301, + "learning_rate": 0.00046883464278700966, + "loss": 3.7336390018463135, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.02107219770550728, + "learning_rate": 0.0004686134171578702, + "loss": 3.722311496734619, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.02068956382572651, + "learning_rate": 0.00046839205809699183, + "loss": 3.767139434814453, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.020190054550766945, + "learning_rate": 0.0004681705657827195, + "loss": 3.6960511207580566, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.018814411014318466, + "learning_rate": 0.0004679489403935058, + "loss": 3.70375919342041, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.017568519338965416, + "learning_rate": 0.00046772718210791005, + "loss": 3.756627321243286, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.018188195303082466, + "learning_rate": 0.00046750529110459903, + "loss": 3.6964683532714844, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.018925746902823448, + "learning_rate": 0.0004672832675623463, + "loss": 3.731172561645508, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.01690756157040596, + "learning_rate": 0.0004670611116600322, + "loss": 3.6955318450927734, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016300756484270096, + "learning_rate": 0.00046683882357664375, + "loss": 3.762632369995117, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.016405068337917328, + "learning_rate": 0.0004666164034912744, + "loss": 3.75130558013916, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.017687877640128136, + "learning_rate": 0.00046639385158312406, + "loss": 3.687530755996704, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.017524344846606255, + "learning_rate": 0.00046617116803149874, + "loss": 3.738790273666382, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.01714804768562317, + "learning_rate": 0.00046594835301581053, + "loss": 3.69827938079834, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018042324110865593, + "learning_rate": 0.00046572540671557745, + "loss": 3.71600341796875, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018648479133844376, + "learning_rate": 0.00046550232931042346, + "loss": 3.742591619491577, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.019198937341570854, + "learning_rate": 0.00046527912098007787, + "loss": 3.739820957183838, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.01796373911201954, + "learning_rate": 0.00046505578190437564, + "loss": 3.719025135040283, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01836620457470417, + "learning_rate": 0.0004648323122632569, + "loss": 3.741434097290039, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.02192627638578415, + "learning_rate": 0.0004646087122367673, + "loss": 3.7553882598876953, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.02147185243666172, + "learning_rate": 0.00046438498200505727, + "loss": 3.7705790996551514, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01855960488319397, + "learning_rate": 0.0004641611217483821, + "loss": 3.704925537109375, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.01925475336611271, + "learning_rate": 0.00046393713164710217, + "loss": 3.711491346359253, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019796503707766533, + "learning_rate": 0.00046371301188168204, + "loss": 3.7241992950439453, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.018511971458792686, + "learning_rate": 0.0004634887626326911, + "loss": 3.6785435676574707, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.018214019015431404, + "learning_rate": 0.00046326438408080293, + "loss": 3.7299962043762207, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.018912190571427345, + "learning_rate": 0.00046303987640679517, + "loss": 3.773698329925537, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.01939854398369789, + "learning_rate": 0.0004628152397915498, + "loss": 3.682321071624756, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.017859861254692078, + "learning_rate": 0.00046259047441605215, + "loss": 3.7254951000213623, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.019249707460403442, + "learning_rate": 0.0004623655804613919, + "loss": 3.7263433933258057, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.019221967086195946, + "learning_rate": 0.00046214055810876194, + "loss": 3.7229225635528564, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01871953159570694, + "learning_rate": 0.00046191540753945886, + "loss": 3.7248311042785645, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01747182011604309, + "learning_rate": 0.00046169012893488204, + "loss": 3.699913740158081, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.019630730152130127, + "learning_rate": 0.00046146472247653484, + "loss": 3.787783145904541, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.072091102600098, + "eval_runtime": 63.1589, + "eval_samples_per_second": 38.664, + "eval_steps_per_second": 1.219, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.022918934002518654, + "learning_rate": 0.0004612391883460228, + "loss": 3.666675567626953, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.0264838095754385, + "learning_rate": 0.00046101352672505493, + "loss": 3.6371259689331055, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.028463594615459442, + "learning_rate": 0.0004607877377954426, + "loss": 3.657130241394043, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.02900792844593525, + "learning_rate": 0.00046056182173909995, + "loss": 3.6365628242492676, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.028593607246875763, + "learning_rate": 0.0004603357787380435, + "loss": 3.647258996963501, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.02878790907561779, + "learning_rate": 0.0004601096089743919, + "loss": 3.682035446166992, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.02980183996260166, + "learning_rate": 0.0004598833126303661, + "loss": 3.623297691345215, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.03375725448131561, + "learning_rate": 0.00045965688988828884, + "loss": 3.6240830421447754, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.035045325756073, + "learning_rate": 0.00045943034093058507, + "loss": 3.6372249126434326, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.03519923985004425, + "learning_rate": 0.000459203665939781, + "loss": 3.6157641410827637, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.03468473628163338, + "learning_rate": 0.00045897686509850487, + "loss": 3.6553335189819336, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.034817393869161606, + "learning_rate": 0.0004587499385894856, + "loss": 3.653806686401367, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.03549059107899666, + "learning_rate": 0.000458522886595554, + "loss": 3.6730189323425293, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.028001410886645317, + "learning_rate": 0.0004582957092996418, + "loss": 3.662490129470825, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.02532520331442356, + "learning_rate": 0.0004580684068847817, + "loss": 3.629936456680298, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.028215918689966202, + "learning_rate": 0.0004578409795341069, + "loss": 3.6507444381713867, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.026823483407497406, + "learning_rate": 0.0004576134274308517, + "loss": 3.661811113357544, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.0281058382242918, + "learning_rate": 0.0004573857507583506, + "loss": 3.6805872917175293, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.02548503689467907, + "learning_rate": 0.00045715794970003874, + "loss": 3.6586413383483887, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.024795135483145714, + "learning_rate": 0.0004569300244394513, + "loss": 3.654989004135132, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.025501688942313194, + "learning_rate": 0.0004567019751602234, + "loss": 3.6915903091430664, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.02377370186150074, + "learning_rate": 0.00045647380204609034, + "loss": 3.660588264465332, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.022150425240397453, + "learning_rate": 0.00045624550528088704, + "loss": 3.64870023727417, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.02415974996984005, + "learning_rate": 0.00045601708504854803, + "loss": 3.6498191356658936, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.02140078693628311, + "learning_rate": 0.0004557885415331074, + "loss": 3.703489065170288, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.022121304646134377, + "learning_rate": 0.00045555987491869865, + "loss": 3.6606979370117188, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.02168990671634674, + "learning_rate": 0.00045533108538955416, + "loss": 3.6801178455352783, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.020228713750839233, + "learning_rate": 0.0004551021731300055, + "loss": 3.6707205772399902, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.018500352278351784, + "learning_rate": 0.0004548731383244834, + "loss": 3.647545576095581, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.020223816856741905, + "learning_rate": 0.0004546439811575169, + "loss": 3.644932985305786, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.019502121955156326, + "learning_rate": 0.00045441470181373375, + "loss": 3.62888240814209, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.01926770992577076, + "learning_rate": 0.0004541853004778603, + "loss": 3.675457715988159, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01888471283018589, + "learning_rate": 0.00045395577733472087, + "loss": 3.636434316635132, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.0201494712382555, + "learning_rate": 0.0004537261325692383, + "loss": 3.679147720336914, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.02178201824426651, + "learning_rate": 0.0004534963663664332, + "loss": 3.671905040740967, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.024073559790849686, + "learning_rate": 0.00045326647891142394, + "loss": 3.6701040267944336, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.02135559171438217, + "learning_rate": 0.0004530364703894268, + "loss": 3.648519992828369, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.019264448434114456, + "learning_rate": 0.0004528063409857554, + "loss": 3.6985023021698, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.01922295056283474, + "learning_rate": 0.0004525760908858209, + "loss": 3.651756763458252, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.01978331431746483, + "learning_rate": 0.00045234572027513174, + "loss": 3.679884910583496, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.01968599483370781, + "learning_rate": 0.00045211522933929317, + "loss": 3.6571812629699707, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.020741639658808708, + "learning_rate": 0.0004518846182640077, + "loss": 3.6505298614501953, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.01842017099261284, + "learning_rate": 0.0004516538872350744, + "loss": 3.6904940605163574, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.018004486337304115, + "learning_rate": 0.0004514230364383893, + "loss": 3.7282047271728516, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.018821008503437042, + "learning_rate": 0.0004511920660599444, + "loss": 3.668164014816284, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.018929854035377502, + "learning_rate": 0.0004509609762858287, + "loss": 3.6672980785369873, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.01979168877005577, + "learning_rate": 0.0004507297673022269, + "loss": 3.6684999465942383, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.01774384267628193, + "learning_rate": 0.0004504984392954199, + "loss": 3.6577749252319336, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01876000687479973, + "learning_rate": 0.00045026699245178444, + "loss": 3.6629252433776855, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.018197249621152878, + "learning_rate": 0.0004500354269577932, + "loss": 3.688983678817749, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.019496265798807144, + "learning_rate": 0.00044980374300001414, + "loss": 3.6602022647857666, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01825535111129284, + "learning_rate": 0.00044957194076511095, + "loss": 3.6889238357543945, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.017113137990236282, + "learning_rate": 0.0004493400204398426, + "loss": 3.684070587158203, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.017993126064538956, + "learning_rate": 0.00044910798221106283, + "loss": 3.675426721572876, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.018354708328843117, + "learning_rate": 0.00044887582626572105, + "loss": 3.675588607788086, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.016362832859158516, + "learning_rate": 0.0004486435527908606, + "loss": 3.6709303855895996, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01772182621061802, + "learning_rate": 0.0004484111619736205, + "loss": 3.6884102821350098, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.019271764904260635, + "learning_rate": 0.0004481786540012336, + "loss": 3.6468968391418457, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.019040528684854507, + "learning_rate": 0.00044794602906102747, + "loss": 3.690387010574341, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.019022610038518906, + "learning_rate": 0.00044771328734042367, + "loss": 3.689873218536377, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.018533391878008842, + "learning_rate": 0.00044748042902693815, + "loss": 3.6877284049987793, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.020825615152716637, + "learning_rate": 0.0004472474543081805, + "loss": 3.662257671356201, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.021880341693758965, + "learning_rate": 0.00044701436337185444, + "loss": 3.650700092315674, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.019974704831838608, + "learning_rate": 0.00044678115640575704, + "loss": 3.684177875518799, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.01903127320110798, + "learning_rate": 0.0004465478335977788, + "loss": 3.658148765563965, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.020999105647206306, + "learning_rate": 0.00044631439513590393, + "loss": 3.68747615814209, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.02138398587703705, + "learning_rate": 0.00044608084120820933, + "loss": 3.656355381011963, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.02216442860662937, + "learning_rate": 0.0004458471720028655, + "loss": 3.6860342025756836, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.019359121099114418, + "learning_rate": 0.0004456133877081352, + "loss": 3.654944658279419, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.020984770730137825, + "learning_rate": 0.0004453794885123744, + "loss": 3.6885194778442383, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.02423972077667713, + "learning_rate": 0.00044514547460403135, + "loss": 3.672712802886963, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.024544453248381615, + "learning_rate": 0.00044491134617164694, + "loss": 3.6484768390655518, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.060352325439453, + "eval_runtime": 57.9556, + "eval_samples_per_second": 42.136, + "eval_steps_per_second": 1.329, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.024072550237178802, + "learning_rate": 0.000444677103403854, + "loss": 3.5796468257904053, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.028701433911919594, + "learning_rate": 0.0004444427464893779, + "loss": 3.604318618774414, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.028119578957557678, + "learning_rate": 0.00044420827561703573, + "loss": 3.608759641647339, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.025335192680358887, + "learning_rate": 0.0004439736909757365, + "loss": 3.5760319232940674, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.0273929201066494, + "learning_rate": 0.0004437389927544805, + "loss": 3.559441089630127, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.026567714288830757, + "learning_rate": 0.0004435041811423603, + "loss": 3.626889228820801, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.027587885037064552, + "learning_rate": 0.0004432692563285591, + "loss": 3.6226272583007812, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.02935059368610382, + "learning_rate": 0.0004430342185023517, + "loss": 3.601921558380127, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.02982175536453724, + "learning_rate": 0.0004427990678531038, + "loss": 3.59737491607666, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.028570258989930153, + "learning_rate": 0.0004425638045702719, + "loss": 3.56337833404541, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.029890350997447968, + "learning_rate": 0.0004423284288434035, + "loss": 3.6294307708740234, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.03206814453005791, + "learning_rate": 0.00044209294086213657, + "loss": 3.590505599975586, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.03169345110654831, + "learning_rate": 0.00044185734081619957, + "loss": 3.581584930419922, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.03313758969306946, + "learning_rate": 0.000441621628895411, + "loss": 3.6088438034057617, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.03317603841423988, + "learning_rate": 0.00044138580528967986, + "loss": 3.6340456008911133, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.034738484770059586, + "learning_rate": 0.0004411498701890051, + "loss": 3.5814096927642822, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.036428093910217285, + "learning_rate": 0.0004409138237834751, + "loss": 3.5939526557922363, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.03293406963348389, + "learning_rate": 0.00044067766626326835, + "loss": 3.577897787094116, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.031030720099806786, + "learning_rate": 0.00044044139781865255, + "loss": 3.6077990531921387, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.03441236540675163, + "learning_rate": 0.00044020501863998516, + "loss": 3.591923713684082, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.03500965237617493, + "learning_rate": 0.00043996852891771236, + "loss": 3.6171927452087402, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.034443050622940063, + "learning_rate": 0.00043973192884236983, + "loss": 3.619724750518799, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.03435051813721657, + "learning_rate": 0.0004394952186045819, + "loss": 3.6443169116973877, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.03230702877044678, + "learning_rate": 0.00043925839839506167, + "loss": 3.5912671089172363, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.03232988342642784, + "learning_rate": 0.0004390214684046111, + "loss": 3.618985414505005, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.029298555105924606, + "learning_rate": 0.00043878442882412015, + "loss": 3.6166677474975586, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.026723450049757957, + "learning_rate": 0.00043854727984456746, + "loss": 3.61362886428833, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.028280874714255333, + "learning_rate": 0.0004383100216570197, + "loss": 3.6410951614379883, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.026700228452682495, + "learning_rate": 0.0004380726544526316, + "loss": 3.5804049968719482, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.023092713207006454, + "learning_rate": 0.00043783517842264534, + "loss": 3.595339298248291, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.024123897776007652, + "learning_rate": 0.0004375975937583914, + "loss": 3.647569417953491, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.02505572699010372, + "learning_rate": 0.0004373599006512871, + "loss": 3.6276652812957764, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.024523509666323662, + "learning_rate": 0.00043712209929283786, + "loss": 3.59525728225708, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.024351581931114197, + "learning_rate": 0.00043688418987463567, + "loss": 3.6081674098968506, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.02530614472925663, + "learning_rate": 0.00043664617258835995, + "loss": 3.6632542610168457, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.023438062518835068, + "learning_rate": 0.0004364080476257769, + "loss": 3.6345431804656982, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.021853657439351082, + "learning_rate": 0.0004361698151787396, + "loss": 3.598644256591797, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.022487422451376915, + "learning_rate": 0.0004359314754391874, + "loss": 3.5966577529907227, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.021353935822844505, + "learning_rate": 0.0004356930285991466, + "loss": 3.6328375339508057, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.021180452778935432, + "learning_rate": 0.00043545447485072926, + "loss": 3.6285812854766846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.020491832867264748, + "learning_rate": 0.000435215814386134, + "loss": 3.592944622039795, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.021603770554065704, + "learning_rate": 0.0004349770473976453, + "loss": 3.613507032394409, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.02146158367395401, + "learning_rate": 0.0004347381740776332, + "loss": 3.6409506797790527, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.02102433145046234, + "learning_rate": 0.0004344991946185538, + "loss": 3.6179330348968506, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.018684551119804382, + "learning_rate": 0.00043426010921294837, + "loss": 3.6385130882263184, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.018416227772831917, + "learning_rate": 0.0004340209180534438, + "loss": 3.642151355743408, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.01915144734084606, + "learning_rate": 0.000433781621332752, + "loss": 3.6089656352996826, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.018299665302038193, + "learning_rate": 0.0004335422192436703, + "loss": 3.6063547134399414, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.018962757661938667, + "learning_rate": 0.00043330271197908023, + "loss": 3.625917434692383, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.017437821254134178, + "learning_rate": 0.00043306309973194874, + "loss": 3.600015640258789, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.017913147807121277, + "learning_rate": 0.000432823382695327, + "loss": 3.6428744792938232, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.018343618139624596, + "learning_rate": 0.0004325835610623508, + "loss": 3.635901927947998, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.020372850820422173, + "learning_rate": 0.00043234363502623994, + "loss": 3.620181083679199, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.02037600241601467, + "learning_rate": 0.0004321036047802985, + "loss": 3.61129093170166, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.021827151998877525, + "learning_rate": 0.00043186347051791466, + "loss": 3.5999555587768555, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.021806657314300537, + "learning_rate": 0.0004316232324325602, + "loss": 3.642509937286377, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02050699107348919, + "learning_rate": 0.0004313828907177906, + "loss": 3.597583293914795, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.020436758175492287, + "learning_rate": 0.0004311424455672448, + "loss": 3.6079912185668945, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.019341666251420975, + "learning_rate": 0.0004309018971746451, + "loss": 3.6196017265319824, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.022348817437887192, + "learning_rate": 0.000430661245733797, + "loss": 3.619025468826294, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.019361121580004692, + "learning_rate": 0.00043042049143858914, + "loss": 3.657182455062866, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.019258935004472733, + "learning_rate": 0.0004301796344829928, + "loss": 3.6408963203430176, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.020622272044420242, + "learning_rate": 0.00042993867506106205, + "loss": 3.6247220039367676, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.019510755315423012, + "learning_rate": 0.0004296976133669336, + "loss": 3.6264584064483643, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.02078665979206562, + "learning_rate": 0.0004294564495948266, + "loss": 3.6364388465881348, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01791800744831562, + "learning_rate": 0.0004292151839390421, + "loss": 3.6029374599456787, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.01922667771577835, + "learning_rate": 0.0004289738165939636, + "loss": 3.582937717437744, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.018473893404006958, + "learning_rate": 0.00042873234775405654, + "loss": 3.6141819953918457, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01846298575401306, + "learning_rate": 0.00042849077761386784, + "loss": 3.638786554336548, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.01876973733305931, + "learning_rate": 0.0004282491063680263, + "loss": 3.617130756378174, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.017995499074459076, + "learning_rate": 0.00042800733421124207, + "loss": 3.645305633544922, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.02105429396033287, + "learning_rate": 0.00042776546133830646, + "loss": 3.6444876194000244, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.049062252044678, + "eval_runtime": 58.0113, + "eval_samples_per_second": 42.095, + "eval_steps_per_second": 1.327, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.02128119394183159, + "learning_rate": 0.00042752348794409227, + "loss": 3.524954080581665, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.028249500319361687, + "learning_rate": 0.00042728141422355287, + "loss": 3.5311479568481445, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.03145391866564751, + "learning_rate": 0.0004270392403717229, + "loss": 3.559330940246582, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.02994118444621563, + "learning_rate": 0.00042679696658371743, + "loss": 3.5257620811462402, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.027448246255517006, + "learning_rate": 0.00042655459305473195, + "loss": 3.5242764949798584, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.0296134352684021, + "learning_rate": 0.0004263121199800425, + "loss": 3.5317091941833496, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.029576804488897324, + "learning_rate": 0.0004260695475550054, + "loss": 3.5340428352355957, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.02941269613802433, + "learning_rate": 0.0004258268759750566, + "loss": 3.5372023582458496, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.03126693144440651, + "learning_rate": 0.0004255841054357124, + "loss": 3.5437135696411133, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.03335367515683174, + "learning_rate": 0.0004253412361325686, + "loss": 3.5598273277282715, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.03028946928679943, + "learning_rate": 0.0004250982682613006, + "loss": 3.5190839767456055, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.028629133477807045, + "learning_rate": 0.00042485520201766315, + "loss": 3.5138821601867676, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.027070026844739914, + "learning_rate": 0.0004246120375974905, + "loss": 3.530958414077759, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.028837760910391808, + "learning_rate": 0.0004243687751966957, + "loss": 3.5233962535858154, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.029884297400712967, + "learning_rate": 0.0004241254150112708, + "loss": 3.5639758110046387, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.024223247542977333, + "learning_rate": 0.000423881957237287, + "loss": 3.5555152893066406, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.026992836967110634, + "learning_rate": 0.0004236384020708935, + "loss": 3.540196418762207, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.02818330004811287, + "learning_rate": 0.0004233947497083185, + "loss": 3.5546071529388428, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.026401042938232422, + "learning_rate": 0.0004231510003458682, + "loss": 3.5556247234344482, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.026020489633083344, + "learning_rate": 0.0004229071541799272, + "loss": 3.568004608154297, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.02803782932460308, + "learning_rate": 0.0004226632114069578, + "loss": 3.5828235149383545, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.02737978659570217, + "learning_rate": 0.0004224191722235004, + "loss": 3.5642499923706055, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.02406371757388115, + "learning_rate": 0.00042217503682617286, + "loss": 3.5530993938446045, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.02230692468583584, + "learning_rate": 0.0004219308054116706, + "loss": 3.5573501586914062, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.021352406591176987, + "learning_rate": 0.0004216864781767667, + "loss": 3.5526974201202393, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.02227999083697796, + "learning_rate": 0.00042144205531831103, + "loss": 3.5866928100585938, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.024190247058868408, + "learning_rate": 0.0004211975370332308, + "loss": 3.517117977142334, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.023196619004011154, + "learning_rate": 0.00042095292351852976, + "loss": 3.4826502799987793, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.02388714998960495, + "learning_rate": 0.0004207082149712888, + "loss": 3.5588626861572266, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.024692339822649956, + "learning_rate": 0.0004204634115886651, + "loss": 3.54828143119812, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.023078206926584244, + "learning_rate": 0.0004202185135678924, + "loss": 3.5679030418395996, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.020665735006332397, + "learning_rate": 0.00041997352110628033, + "loss": 3.5500283241271973, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.02022322081029415, + "learning_rate": 0.00041972843440121527, + "loss": 3.592104911804199, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.021137656643986702, + "learning_rate": 0.00041948325365015894, + "loss": 3.583833694458008, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.021026931703090668, + "learning_rate": 0.0004192379790506491, + "loss": 3.556403875350952, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.020423648878932, + "learning_rate": 0.00041899261080029904, + "loss": 3.5839173793792725, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.020935002714395523, + "learning_rate": 0.00041874714909679754, + "loss": 3.5794217586517334, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.020799759775400162, + "learning_rate": 0.00041850159413790863, + "loss": 3.6033859252929688, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.019480884075164795, + "learning_rate": 0.0004182559461214715, + "loss": 3.580097198486328, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.019300777465105057, + "learning_rate": 0.0004180102052454004, + "loss": 3.550474166870117, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.020265772938728333, + "learning_rate": 0.00041776437170768427, + "loss": 3.5711655616760254, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.019972780719399452, + "learning_rate": 0.0004175184457063868, + "loss": 3.5636885166168213, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.019635623320937157, + "learning_rate": 0.000417272427439646, + "loss": 3.5851311683654785, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.021451901644468307, + "learning_rate": 0.00041702631710567443, + "loss": 3.5956759452819824, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.019456414505839348, + "learning_rate": 0.00041678011490275875, + "loss": 3.595731258392334, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.019757689908146858, + "learning_rate": 0.00041653382102925957, + "loss": 3.569652557373047, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.020571604371070862, + "learning_rate": 0.00041628743568361147, + "loss": 3.6007747650146484, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.02153848111629486, + "learning_rate": 0.00041604095906432265, + "loss": 3.5915260314941406, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.020875850692391396, + "learning_rate": 0.00041579439136997493, + "loss": 3.5582690238952637, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.02089419960975647, + "learning_rate": 0.00041554773279922337, + "loss": 3.5756618976593018, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01961098238825798, + "learning_rate": 0.0004153009835507963, + "loss": 3.5918776988983154, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.021075716242194176, + "learning_rate": 0.0004150541438234952, + "loss": 3.5342202186584473, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.020745476707816124, + "learning_rate": 0.00041480721381619435, + "loss": 3.5976028442382812, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.02001352794468403, + "learning_rate": 0.00041456019372784086, + "loss": 3.5686182975769043, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.021684160456061363, + "learning_rate": 0.0004143130837574543, + "loss": 3.553445339202881, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.022551920264959335, + "learning_rate": 0.0004140658841041267, + "loss": 3.5712709426879883, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.022758224979043007, + "learning_rate": 0.0004138185949670224, + "loss": 3.6101667881011963, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.020714636892080307, + "learning_rate": 0.00041357121654537766, + "loss": 3.5740506649017334, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.022694187238812447, + "learning_rate": 0.00041332374903850104, + "loss": 3.5663979053497314, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.021664340049028397, + "learning_rate": 0.00041307619264577234, + "loss": 3.5830888748168945, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.019072270020842552, + "learning_rate": 0.0004128285475666436, + "loss": 3.5861973762512207, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.020153764635324478, + "learning_rate": 0.0004125808140006378, + "loss": 3.566444158554077, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019666558131575584, + "learning_rate": 0.0004123329921473494, + "loss": 3.5789592266082764, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.02101624570786953, + "learning_rate": 0.0004120850822064439, + "loss": 3.551722526550293, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.01951020397245884, + "learning_rate": 0.00041183708437765815, + "loss": 3.6069014072418213, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.01985476352274418, + "learning_rate": 0.00041158899886079925, + "loss": 3.5944924354553223, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02195335365831852, + "learning_rate": 0.00041134082585574534, + "loss": 3.5828046798706055, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02105172537267208, + "learning_rate": 0.00041109256556244497, + "loss": 3.5830953121185303, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.01888827420771122, + "learning_rate": 0.00041084421818091685, + "loss": 3.5656847953796387, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.017867228016257286, + "learning_rate": 0.0004105957839112501, + "loss": 3.5792243480682373, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.018577905371785164, + "learning_rate": 0.0004103472629536036, + "loss": 3.56992506980896, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.020045632496476173, + "learning_rate": 0.0004100986555082063, + "loss": 3.590876579284668, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.044339656829834, + "eval_runtime": 57.1435, + "eval_samples_per_second": 42.735, + "eval_steps_per_second": 1.347, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.024135645478963852, + "learning_rate": 0.0004098499617753567, + "loss": 3.4856467247009277, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.031053155660629272, + "learning_rate": 0.00040960118195542273, + "loss": 3.455773115158081, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.035190410912036896, + "learning_rate": 0.00040935231624884204, + "loss": 3.506619930267334, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.03611094877123833, + "learning_rate": 0.0004091033648561211, + "loss": 3.5227696895599365, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.034121073782444, + "learning_rate": 0.00040885432797783543, + "loss": 3.4970145225524902, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.03629382699728012, + "learning_rate": 0.0004086052058146297, + "loss": 3.506425619125366, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.03709045797586441, + "learning_rate": 0.00040835599856721725, + "loss": 3.499274969100952, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.034826792776584625, + "learning_rate": 0.00040810670643637954, + "loss": 3.5029163360595703, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.03344690054655075, + "learning_rate": 0.000407857329622967, + "loss": 3.4816911220550537, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.03136949986219406, + "learning_rate": 0.00040760786832789786, + "loss": 3.5154659748077393, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.030533097684383392, + "learning_rate": 0.0004073583227521588, + "loss": 3.509666919708252, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.03298318013548851, + "learning_rate": 0.0004071086930968039, + "loss": 3.5096616744995117, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.031186232343316078, + "learning_rate": 0.00040685897956295545, + "loss": 3.4752590656280518, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.028008371591567993, + "learning_rate": 0.0004066091823518031, + "loss": 3.498826503753662, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.028206519782543182, + "learning_rate": 0.00040635930166460385, + "loss": 3.5281882286071777, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.024417119100689888, + "learning_rate": 0.00040610933770268226, + "loss": 3.506791830062866, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.025321070104837418, + "learning_rate": 0.00040585929066742964, + "loss": 3.494256019592285, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.023430505767464638, + "learning_rate": 0.00040560916076030435, + "loss": 3.533010482788086, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.023371374234557152, + "learning_rate": 0.00040535894818283154, + "loss": 3.5197086334228516, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.02414027415215969, + "learning_rate": 0.0004051086531366031, + "loss": 3.4707889556884766, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.026652060449123383, + "learning_rate": 0.000404858275823277, + "loss": 3.498159885406494, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.024486590176820755, + "learning_rate": 0.000404607816444578, + "loss": 3.50034236907959, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.023153992369771004, + "learning_rate": 0.00040435727520229644, + "loss": 3.499504804611206, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.02222427912056446, + "learning_rate": 0.0004041066522982891, + "loss": 3.4953579902648926, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.023033641278743744, + "learning_rate": 0.000403855947934478, + "loss": 3.5181496143341064, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.02141181193292141, + "learning_rate": 0.00040360516231285155, + "loss": 3.552860736846924, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.023448696359992027, + "learning_rate": 0.00040335429563546286, + "loss": 3.5153324604034424, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.022410666570067406, + "learning_rate": 0.0004031033481044308, + "loss": 3.4963512420654297, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.020326802507042885, + "learning_rate": 0.00040285231992193924, + "loss": 3.528319835662842, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021117394790053368, + "learning_rate": 0.0004026012112902372, + "loss": 3.5192837715148926, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.021317072212696075, + "learning_rate": 0.0004023500224116381, + "loss": 3.5081021785736084, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.021116144955158234, + "learning_rate": 0.00040209875348852037, + "loss": 3.5167174339294434, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019834019243717194, + "learning_rate": 0.00040184740472332705, + "loss": 3.511960983276367, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.02089378982782364, + "learning_rate": 0.000401595976318565, + "loss": 3.493330955505371, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.021129710599780083, + "learning_rate": 0.00040134446847680584, + "loss": 3.544156551361084, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.021163299679756165, + "learning_rate": 0.0004010928814006846, + "loss": 3.5261003971099854, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.02105306088924408, + "learning_rate": 0.0004008412152929007, + "loss": 3.543973445892334, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.02150859497487545, + "learning_rate": 0.0004005894703562168, + "loss": 3.487217426300049, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.021666452288627625, + "learning_rate": 0.0004003376467934593, + "loss": 3.4933671951293945, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.02180723287165165, + "learning_rate": 0.0004000857448075178, + "loss": 3.55198073387146, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.02041625790297985, + "learning_rate": 0.0003998337646013455, + "loss": 3.529571533203125, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.022977063432335854, + "learning_rate": 0.00039958170637795806, + "loss": 3.526432752609253, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.022725161164999008, + "learning_rate": 0.0003993295703404341, + "loss": 3.5258829593658447, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.021337240934371948, + "learning_rate": 0.00039907735669191544, + "loss": 3.4897351264953613, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.020783131942152977, + "learning_rate": 0.00039882506563560573, + "loss": 3.5497589111328125, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.022259317338466644, + "learning_rate": 0.0003985726973747715, + "loss": 3.5225064754486084, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.020283272489905357, + "learning_rate": 0.00039832025211274113, + "loss": 3.5453078746795654, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.020656265318393707, + "learning_rate": 0.00039806773005290544, + "loss": 3.533703327178955, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.02104976586997509, + "learning_rate": 0.0003978151313987168, + "loss": 3.4885146617889404, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.021411150693893433, + "learning_rate": 0.00039756245635368933, + "loss": 3.533268690109253, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.021906673908233643, + "learning_rate": 0.00039730970512139876, + "loss": 3.5483224391937256, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.02415728010237217, + "learning_rate": 0.0003970568779054822, + "loss": 3.5084290504455566, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.023435061797499657, + "learning_rate": 0.000396803974909638, + "loss": 3.5498428344726562, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.02183385379612446, + "learning_rate": 0.0003965509963376255, + "loss": 3.484086751937866, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.022856134921312332, + "learning_rate": 0.000396297942393265, + "loss": 3.5339231491088867, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.021679896861314774, + "learning_rate": 0.0003960448132804375, + "loss": 3.547191619873047, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.021262124180793762, + "learning_rate": 0.0003957916092030845, + "loss": 3.5263822078704834, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.02265402302145958, + "learning_rate": 0.00039553833036520803, + "loss": 3.5361135005950928, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.020080721005797386, + "learning_rate": 0.0003952849769708702, + "loss": 3.5247459411621094, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.02187187783420086, + "learning_rate": 0.0003950315492241932, + "loss": 3.54129695892334, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.02417500875890255, + "learning_rate": 0.0003947780473293593, + "loss": 3.598796844482422, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.02416270412504673, + "learning_rate": 0.00039452447149061054, + "loss": 3.5207936763763428, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.021958643570542336, + "learning_rate": 0.0003942708219122481, + "loss": 3.5395452976226807, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.021813420578837395, + "learning_rate": 0.000394017098798633, + "loss": 3.5469255447387695, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020607495680451393, + "learning_rate": 0.0003937633023541854, + "loss": 3.554851531982422, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.02091527171432972, + "learning_rate": 0.0003935094327833845, + "loss": 3.492576837539673, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.021089257672429085, + "learning_rate": 0.0003932554902907683, + "loss": 3.5154709815979004, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.019659532234072685, + "learning_rate": 0.0003930014750809338, + "loss": 3.5444283485412598, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01896163448691368, + "learning_rate": 0.0003927473873585365, + "loss": 3.589625358581543, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020288823172450066, + "learning_rate": 0.0003924932273282903, + "loss": 3.522411823272705, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.020222362130880356, + "learning_rate": 0.00039223899519496723, + "loss": 3.5244359970092773, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.022349944338202477, + "learning_rate": 0.00039198469116339757, + "loss": 3.5362396240234375, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.047028064727783, + "eval_runtime": 62.8948, + "eval_samples_per_second": 38.827, + "eval_steps_per_second": 1.224, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.02202913723886013, + "learning_rate": 0.0003917303154384694, + "loss": 3.4249138832092285, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.03153969347476959, + "learning_rate": 0.00039147586822512885, + "loss": 3.4436073303222656, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.03747870773077011, + "learning_rate": 0.0003912213497283793, + "loss": 3.4879579544067383, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.036144088953733444, + "learning_rate": 0.0003909667601532819, + "loss": 3.419151782989502, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.03983336314558983, + "learning_rate": 0.00039071209970495445, + "loss": 3.4331436157226562, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.045912645757198334, + "learning_rate": 0.00039045736858857273, + "loss": 3.4304769039154053, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.04005753993988037, + "learning_rate": 0.0003902025670093687, + "loss": 3.478600025177002, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.03488151356577873, + "learning_rate": 0.00038994769517263156, + "loss": 3.4650232791900635, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.03603382781147957, + "learning_rate": 0.0003896927532837069, + "loss": 3.449807643890381, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.03817906603217125, + "learning_rate": 0.0003894377415479967, + "loss": 3.4556589126586914, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.03523671627044678, + "learning_rate": 0.00038918266017095936, + "loss": 3.477426767349243, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.03284068778157234, + "learning_rate": 0.0003889275093581096, + "loss": 3.478672981262207, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.02992352657020092, + "learning_rate": 0.00038867228931501737, + "loss": 3.4810047149658203, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.02858242578804493, + "learning_rate": 0.0003884170002473091, + "loss": 3.467811107635498, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.027912171557545662, + "learning_rate": 0.0003881616423606666, + "loss": 3.4326171875, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.028351375833153725, + "learning_rate": 0.00038790621586082697, + "loss": 3.4533729553222656, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.027947887778282166, + "learning_rate": 0.0003876507209535829, + "loss": 3.420295476913452, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.028510035946965218, + "learning_rate": 0.00038739515784478186, + "loss": 3.438023328781128, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.027707934379577637, + "learning_rate": 0.00038713952674032654, + "loss": 3.4844181537628174, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.0262764822691679, + "learning_rate": 0.0003868838278461743, + "loss": 3.4363749027252197, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.02713681012392044, + "learning_rate": 0.0003866280613683371, + "loss": 3.4501900672912598, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.0299485195428133, + "learning_rate": 0.00038637222751288126, + "loss": 3.461942672729492, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.027047986164689064, + "learning_rate": 0.0003861163264859277, + "loss": 3.454293727874756, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.02436736971139908, + "learning_rate": 0.0003858603584936511, + "loss": 3.4663405418395996, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.02601189725100994, + "learning_rate": 0.0003856043237422803, + "loss": 3.502030849456787, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.02463468722999096, + "learning_rate": 0.0003853482224380978, + "loss": 3.462491035461426, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.021793190389871597, + "learning_rate": 0.00038509205478744, + "loss": 3.4476377964019775, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.023809107020497322, + "learning_rate": 0.0003848358209966963, + "loss": 3.4701178073883057, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.02557182125747204, + "learning_rate": 0.0003845795212723099, + "loss": 3.4932539463043213, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.023424552753567696, + "learning_rate": 0.00038432315582077664, + "loss": 3.4779105186462402, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.02195148915052414, + "learning_rate": 0.0003840667248486456, + "loss": 3.470506191253662, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.022897250950336456, + "learning_rate": 0.0003838102285625186, + "loss": 3.48982572555542, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.023033805191516876, + "learning_rate": 0.00038355366716905006, + "loss": 3.4845101833343506, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.022774485871195793, + "learning_rate": 0.0003832970408749467, + "loss": 3.5060324668884277, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.02193611115217209, + "learning_rate": 0.00038304034988696816, + "loss": 3.511518955230713, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.02130570448935032, + "learning_rate": 0.00038278359441192516, + "loss": 3.468502998352051, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.020402755588293076, + "learning_rate": 0.00038252677465668136, + "loss": 3.4786908626556396, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.022664198651909828, + "learning_rate": 0.00038226989082815156, + "loss": 3.479928970336914, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.02454056777060032, + "learning_rate": 0.00038201294313330245, + "loss": 3.479947566986084, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.022096732631325722, + "learning_rate": 0.00038175593177915227, + "loss": 3.49478816986084, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.021499406546354294, + "learning_rate": 0.0003814988569727704, + "loss": 3.494933605194092, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.022325703874230385, + "learning_rate": 0.0003812417189212773, + "loss": 3.5041463375091553, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.020674455910921097, + "learning_rate": 0.00038098451783184447, + "loss": 3.472404718399048, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.02131054364144802, + "learning_rate": 0.00038072725391169416, + "loss": 3.497795343399048, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.02126218006014824, + "learning_rate": 0.0003804699273680994, + "loss": 3.483281373977661, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.0217830128967762, + "learning_rate": 0.0003802125384083834, + "loss": 3.4704298973083496, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.02156917378306389, + "learning_rate": 0.0003799550872399197, + "loss": 3.4994399547576904, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.02047055773437023, + "learning_rate": 0.0003796975740701323, + "loss": 3.4316084384918213, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.019766805693507195, + "learning_rate": 0.0003794399991064948, + "loss": 3.452223777770996, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.020028885453939438, + "learning_rate": 0.00037918236255653074, + "loss": 3.485278844833374, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.020312055945396423, + "learning_rate": 0.00037892466462781306, + "loss": 3.4789156913757324, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.020725280046463013, + "learning_rate": 0.00037866690552796446, + "loss": 3.495528221130371, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.02088077738881111, + "learning_rate": 0.00037840908546465696, + "loss": 3.4972569942474365, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.021010037511587143, + "learning_rate": 0.0003781512046456111, + "loss": 3.473562240600586, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.019641095772385597, + "learning_rate": 0.0003778932632785972, + "loss": 3.4782915115356445, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.02063329517841339, + "learning_rate": 0.00037763526157143376, + "loss": 3.4681782722473145, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.0213928259909153, + "learning_rate": 0.00037737719973198825, + "loss": 3.472263813018799, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.02135022170841694, + "learning_rate": 0.0003771190779681762, + "loss": 3.502300262451172, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.02053583413362503, + "learning_rate": 0.0003768608964879619, + "loss": 3.489243745803833, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.021575238555669785, + "learning_rate": 0.00037660265549935724, + "loss": 3.491074562072754, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.022335771471261978, + "learning_rate": 0.00037634435521042257, + "loss": 3.5298843383789062, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.021198464557528496, + "learning_rate": 0.0003760859958292656, + "loss": 3.5156898498535156, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.020986037328839302, + "learning_rate": 0.000375827577564042, + "loss": 3.4574475288391113, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.022639811038970947, + "learning_rate": 0.0003755691006229545, + "loss": 3.4952611923217773, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.019711066037416458, + "learning_rate": 0.0003753105652142534, + "loss": 3.531296730041504, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.02058061584830284, + "learning_rate": 0.000375051971546236, + "loss": 3.5340657234191895, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.019849255681037903, + "learning_rate": 0.0003747933198272465, + "loss": 3.5028295516967773, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01954430714249611, + "learning_rate": 0.00037453461026567604, + "loss": 3.51088809967041, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.019234873354434967, + "learning_rate": 0.00037427584306996196, + "loss": 3.472560167312622, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.021085912361741066, + "learning_rate": 0.0003740170184485888, + "loss": 3.468876838684082, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.020300406962633133, + "learning_rate": 0.0003737581366100864, + "loss": 3.4903886318206787, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02380351349711418, + "learning_rate": 0.0003734991977630315, + "loss": 3.5069704055786133, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.048111915588379, + "eval_runtime": 58.312, + "eval_samples_per_second": 41.878, + "eval_steps_per_second": 1.32, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02379327267408371, + "learning_rate": 0.0003732402021160463, + "loss": 3.396376132965088, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.03204222396016121, + "learning_rate": 0.00037298114987779885, + "loss": 3.4102582931518555, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.034993767738342285, + "learning_rate": 0.000372722041257003, + "loss": 3.385697364807129, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.03387003391981125, + "learning_rate": 0.00037246287646241783, + "loss": 3.4120078086853027, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.03282478079199791, + "learning_rate": 0.0003722036557028478, + "loss": 3.404122829437256, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.03053346462547779, + "learning_rate": 0.0003719443791871422, + "loss": 3.368051767349243, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.029722390696406364, + "learning_rate": 0.0003716850471241958, + "loss": 3.4066247940063477, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.0287226140499115, + "learning_rate": 0.0003714256597229474, + "loss": 3.4212646484375, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.028156550601124763, + "learning_rate": 0.0003711662171923809, + "loss": 3.4192593097686768, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.03021939843893051, + "learning_rate": 0.0003709067197415244, + "loss": 3.3973188400268555, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.03314025700092316, + "learning_rate": 0.00037064716757945036, + "loss": 3.422978401184082, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.03156686946749687, + "learning_rate": 0.0003703875609152753, + "loss": 3.4052343368530273, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.03357478231191635, + "learning_rate": 0.0003701278999581595, + "loss": 3.383730411529541, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.03261081501841545, + "learning_rate": 0.0003698681849173073, + "loss": 3.421250820159912, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.03252755105495453, + "learning_rate": 0.0003696084160019662, + "loss": 3.39884352684021, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.03177274018526077, + "learning_rate": 0.0003693485934214274, + "loss": 3.4175422191619873, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.028845487162470818, + "learning_rate": 0.0003690887173850253, + "loss": 3.3908143043518066, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.026045162230730057, + "learning_rate": 0.0003688287881021374, + "loss": 3.4417710304260254, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.027937524020671844, + "learning_rate": 0.00036856880578218376, + "loss": 3.410548210144043, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.025780271738767624, + "learning_rate": 0.00036830877063462786, + "loss": 3.4069018363952637, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.026093080639839172, + "learning_rate": 0.0003680486828689749, + "loss": 3.4199600219726562, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.025765180587768555, + "learning_rate": 0.00036778854269477315, + "loss": 3.442903995513916, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.022773101925849915, + "learning_rate": 0.0003675283503216127, + "loss": 3.394232749938965, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.025135623291134834, + "learning_rate": 0.000367268105959126, + "loss": 3.4466171264648438, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.02491890639066696, + "learning_rate": 0.00036700780981698705, + "loss": 3.418689727783203, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.02521841786801815, + "learning_rate": 0.0003667474621049119, + "loss": 3.4213061332702637, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.025186199694871902, + "learning_rate": 0.00036648706303265795, + "loss": 3.437631130218506, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.022097431123256683, + "learning_rate": 0.000366226612810024, + "loss": 3.4014902114868164, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.024009618908166885, + "learning_rate": 0.00036596611164685025, + "loss": 3.4292187690734863, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.022758696228265762, + "learning_rate": 0.0003657055597530175, + "loss": 3.4105381965637207, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.02447321265935898, + "learning_rate": 0.00036544495733844797, + "loss": 3.4196958541870117, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.025843804702162743, + "learning_rate": 0.000365184304613104, + "loss": 3.4343783855438232, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.022433951497077942, + "learning_rate": 0.0003649236017869892, + "loss": 3.440580368041992, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.02250702679157257, + "learning_rate": 0.00036466284907014685, + "loss": 3.4255261421203613, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.022953296080231667, + "learning_rate": 0.0003644020466726608, + "loss": 3.442258358001709, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.021977661177515984, + "learning_rate": 0.00036414119480465473, + "loss": 3.4058687686920166, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.021751178428530693, + "learning_rate": 0.0003638802936762925, + "loss": 3.4211649894714355, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.022764205932617188, + "learning_rate": 0.0003636193434977771, + "loss": 3.424694061279297, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.024827787652611732, + "learning_rate": 0.0003633583444793516, + "loss": 3.458076238632202, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.021888602524995804, + "learning_rate": 0.00036309729683129814, + "loss": 3.4509153366088867, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.021961702033877373, + "learning_rate": 0.00036283620076393814, + "loss": 3.4147253036499023, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.02235778421163559, + "learning_rate": 0.000362575056487632, + "loss": 3.4392640590667725, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.021714260801672935, + "learning_rate": 0.00036231386421277877, + "loss": 3.444026470184326, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.020996905863285065, + "learning_rate": 0.0003620526241498167, + "loss": 3.4168434143066406, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.020528919994831085, + "learning_rate": 0.0003617913365092218, + "loss": 3.4831130504608154, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.021056141704320908, + "learning_rate": 0.0003615300015015091, + "loss": 3.430905342102051, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.021655110642313957, + "learning_rate": 0.0003612686193372312, + "loss": 3.4588241577148438, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.02344612590968609, + "learning_rate": 0.00036100719022697924, + "loss": 3.4202680587768555, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.02264723740518093, + "learning_rate": 0.0003607457143813819, + "loss": 3.4669015407562256, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.02377854660153389, + "learning_rate": 0.00036048419201110553, + "loss": 3.5059216022491455, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.02299923449754715, + "learning_rate": 0.00036022262332685383, + "loss": 3.4796524047851562, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.021110547706484795, + "learning_rate": 0.0003599610085393681, + "loss": 3.4342408180236816, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.022060122340917587, + "learning_rate": 0.0003596993478594267, + "loss": 3.426923990249634, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.021077830344438553, + "learning_rate": 0.0003594376414978447, + "loss": 3.4636270999908447, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.02209082990884781, + "learning_rate": 0.0003591758896654745, + "loss": 3.441979169845581, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.022408071905374527, + "learning_rate": 0.0003589140925732045, + "loss": 3.490835666656494, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.022462666034698486, + "learning_rate": 0.0003586522504319602, + "loss": 3.452770233154297, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.021295800805091858, + "learning_rate": 0.0003583903634527029, + "loss": 3.4377408027648926, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.020457757636904716, + "learning_rate": 0.0003581284318464302, + "loss": 3.4191250801086426, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.021163633093237877, + "learning_rate": 0.00035786645582417564, + "loss": 3.4797091484069824, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.021495461463928223, + "learning_rate": 0.00035760443559700863, + "loss": 3.4763145446777344, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.021050071343779564, + "learning_rate": 0.00035734237137603417, + "loss": 3.4687209129333496, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.02150929905474186, + "learning_rate": 0.0003570802633723927, + "loss": 3.4819254875183105, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.0212967898696661, + "learning_rate": 0.0003568181117972597, + "loss": 3.490407705307007, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.02123691327869892, + "learning_rate": 0.00035655591686184626, + "loss": 3.4418535232543945, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.02193271741271019, + "learning_rate": 0.0003562936787773979, + "loss": 3.446892738342285, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.02065582573413849, + "learning_rate": 0.000356031397755195, + "loss": 3.4308691024780273, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.019487040117383003, + "learning_rate": 0.00035576907400655303, + "loss": 3.487239122390747, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.021190891042351723, + "learning_rate": 0.00035550670774282106, + "loss": 3.4635558128356934, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.0209872554987669, + "learning_rate": 0.0003552442991753831, + "loss": 3.443793773651123, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.02035645768046379, + "learning_rate": 0.00035498184851565697, + "loss": 3.473198652267456, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.024002201855182648, + "learning_rate": 0.0003547193559750944, + "loss": 3.4244751930236816, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.057085990905762, + "eval_runtime": 62.4735, + "eval_samples_per_second": 39.089, + "eval_steps_per_second": 1.233, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.02450242079794407, + "learning_rate": 0.00035445682176518087, + "loss": 3.361237049102783, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.030813416466116905, + "learning_rate": 0.00035419424609743546, + "loss": 3.365708351135254, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.03279578685760498, + "learning_rate": 0.00035393162918341065, + "loss": 3.325385570526123, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.03293970227241516, + "learning_rate": 0.00035366897123469213, + "loss": 3.356811285018921, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.033190157264471054, + "learning_rate": 0.0003534062724628986, + "loss": 3.406217098236084, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.033370476216077805, + "learning_rate": 0.00035314353307968195, + "loss": 3.3579213619232178, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.03351330757141113, + "learning_rate": 0.00035288075329672634, + "loss": 3.3755106925964355, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.034596871584653854, + "learning_rate": 0.0003526179333257487, + "loss": 3.3576536178588867, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.03177076578140259, + "learning_rate": 0.0003523550733784984, + "loss": 3.3123257160186768, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.0289607010781765, + "learning_rate": 0.00035209217366675706, + "loss": 3.370959997177124, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.03265267238020897, + "learning_rate": 0.000351829234402338, + "loss": 3.332608938217163, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.03397805988788605, + "learning_rate": 0.0003515662557970867, + "loss": 3.364213228225708, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.03100395016372204, + "learning_rate": 0.0003513032380628804, + "loss": 3.388005256652832, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.03224753588438034, + "learning_rate": 0.0003510401814116277, + "loss": 3.3267595767974854, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.03034576214849949, + "learning_rate": 0.0003507770860552684, + "loss": 3.3954129219055176, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.026316817849874496, + "learning_rate": 0.00035051395220577397, + "loss": 3.3543832302093506, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.027714375406503677, + "learning_rate": 0.0003502507800751464, + "loss": 3.4190099239349365, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.028731998056173325, + "learning_rate": 0.0003499875698754187, + "loss": 3.342031955718994, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02845810353755951, + "learning_rate": 0.00034972432181865467, + "loss": 3.356130599975586, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.02842753194272518, + "learning_rate": 0.00034946103611694854, + "loss": 3.3709869384765625, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.02942183054983616, + "learning_rate": 0.00034919771298242477, + "loss": 3.373588800430298, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.02958574891090393, + "learning_rate": 0.00034893435262723814, + "loss": 3.377514600753784, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.029524607583880424, + "learning_rate": 0.00034867095526357325, + "loss": 3.398663282394409, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.028092598542571068, + "learning_rate": 0.0003484075211036446, + "loss": 3.3742518424987793, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.026652829721570015, + "learning_rate": 0.0003481440503596964, + "loss": 3.387667179107666, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.024675756692886353, + "learning_rate": 0.0003478805432440021, + "loss": 3.3737070560455322, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.02491692826151848, + "learning_rate": 0.0003476169999688648, + "loss": 3.359217643737793, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.024601681157946587, + "learning_rate": 0.00034735342074661654, + "loss": 3.3902688026428223, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.02313913404941559, + "learning_rate": 0.0003470898057896183, + "loss": 3.3700339794158936, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.02418898604810238, + "learning_rate": 0.0003468261553102599, + "loss": 3.4265692234039307, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.023382030427455902, + "learning_rate": 0.00034656246952095984, + "loss": 3.378389835357666, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.02340470254421234, + "learning_rate": 0.0003462987486341648, + "loss": 3.392937421798706, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.022166339680552483, + "learning_rate": 0.00034603499286235006, + "loss": 3.39414119720459, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.022175129503011703, + "learning_rate": 0.0003457712024180188, + "loss": 3.407447099685669, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.02368207648396492, + "learning_rate": 0.0003455073775137025, + "loss": 3.415076971054077, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.02263982594013214, + "learning_rate": 0.0003452435183619598, + "loss": 3.383986473083496, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.02147712931036949, + "learning_rate": 0.0003449796251753773, + "loss": 3.403963565826416, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.022987166419625282, + "learning_rate": 0.00034471569816656915, + "loss": 3.3883113861083984, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.02240179106593132, + "learning_rate": 0.00034445173754817646, + "loss": 3.4349911212921143, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.02206592820584774, + "learning_rate": 0.00034418774353286747, + "loss": 3.426473617553711, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.022650036960840225, + "learning_rate": 0.0003439237163333375, + "loss": 3.4251773357391357, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.023399339988827705, + "learning_rate": 0.0003436596561623084, + "loss": 3.371610641479492, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.02192695252597332, + "learning_rate": 0.0003433955632325288, + "loss": 3.387253761291504, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.02241935208439827, + "learning_rate": 0.00034313143775677353, + "loss": 3.434523105621338, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.023202672600746155, + "learning_rate": 0.00034286727994784367, + "loss": 3.369771957397461, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.021333014592528343, + "learning_rate": 0.0003426030900185665, + "loss": 3.439004421234131, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.023988714441657066, + "learning_rate": 0.0003423388681817949, + "loss": 3.396838665008545, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.023885777220129967, + "learning_rate": 0.00034207461465040793, + "loss": 3.4114322662353516, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.02263713628053665, + "learning_rate": 0.0003418103296373096, + "loss": 3.4139351844787598, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.022404303774237633, + "learning_rate": 0.00034154601335542964, + "loss": 3.4200823307037354, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.021705076098442078, + "learning_rate": 0.00034128166601772304, + "loss": 3.4092609882354736, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.02273249812424183, + "learning_rate": 0.0003410172878371695, + "loss": 3.3931846618652344, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.022552402690052986, + "learning_rate": 0.00034075287902677394, + "loss": 3.396566867828369, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.023219924420118332, + "learning_rate": 0.0003404884397995655, + "loss": 3.369025707244873, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.022560900077223778, + "learning_rate": 0.00034022397036859837, + "loss": 3.4248225688934326, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.023370858281850815, + "learning_rate": 0.0003399594709469506, + "loss": 3.431628942489624, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.02469109185039997, + "learning_rate": 0.00033969494174772465, + "loss": 3.406843662261963, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.02417670376598835, + "learning_rate": 0.0003394303829840469, + "loss": 3.445046901702881, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.0237257182598114, + "learning_rate": 0.00033916579486906764, + "loss": 3.4490597248077393, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.02295520156621933, + "learning_rate": 0.0003389011776159607, + "loss": 3.4009876251220703, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.023158475756645203, + "learning_rate": 0.0003386365314379233, + "loss": 3.4383902549743652, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.023618478327989578, + "learning_rate": 0.0003383718565481763, + "loss": 3.421168804168701, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.022055329754948616, + "learning_rate": 0.0003381071531599633, + "loss": 3.382946729660034, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.021441346034407616, + "learning_rate": 0.00033784242148655115, + "loss": 3.4464712142944336, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.023031961172819138, + "learning_rate": 0.00033757766174122934, + "loss": 3.42844295501709, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.022832952439785004, + "learning_rate": 0.00033731287413731005, + "loss": 3.398231029510498, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.022094765678048134, + "learning_rate": 0.0003370480588881278, + "loss": 3.4270074367523193, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.0218376275151968, + "learning_rate": 0.0003367832162070395, + "loss": 3.4681742191314697, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.02364739589393139, + "learning_rate": 0.000336518346307424, + "loss": 3.431920051574707, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.023402225226163864, + "learning_rate": 0.0003362534494026824, + "loss": 3.4189023971557617, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.022775836288928986, + "learning_rate": 0.0003359885257062372, + "loss": 3.409396171569824, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.02445383556187153, + "learning_rate": 0.0003357235754315328, + "loss": 3.41017746925354, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.066352367401123, + "eval_runtime": 58.1658, + "eval_samples_per_second": 41.983, + "eval_steps_per_second": 1.324, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.02710847556591034, + "learning_rate": 0.0003354585987920345, + "loss": 3.3155131340026855, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.03428799286484718, + "learning_rate": 0.0003351935960012296, + "loss": 3.327408790588379, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.03434379771351814, + "learning_rate": 0.0003349285672726259, + "loss": 3.3282251358032227, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.03405233472585678, + "learning_rate": 0.0003346635128197522, + "loss": 3.2932448387145996, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.031194346025586128, + "learning_rate": 0.0003343984328561581, + "loss": 3.336615800857544, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.031816478818655014, + "learning_rate": 0.00033413332759541376, + "loss": 3.3375275135040283, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.03413299471139908, + "learning_rate": 0.0003338681972511098, + "loss": 3.3710694313049316, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.034200165420770645, + "learning_rate": 0.0003336030420368568, + "loss": 3.3248214721679688, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.030694160610437393, + "learning_rate": 0.0003333378621662857, + "loss": 3.3297901153564453, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.031172804534435272, + "learning_rate": 0.00033307265785304684, + "loss": 3.287217140197754, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.031914614140987396, + "learning_rate": 0.00033280742931081076, + "loss": 3.3218531608581543, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.0332055389881134, + "learning_rate": 0.00033254217675326737, + "loss": 3.3454785346984863, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.034956980496644974, + "learning_rate": 0.0003322769003941257, + "loss": 3.3193702697753906, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.03618055582046509, + "learning_rate": 0.00033201160044711423, + "loss": 3.3661017417907715, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.030322827398777008, + "learning_rate": 0.0003317462771259802, + "loss": 3.353989601135254, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.029768960550427437, + "learning_rate": 0.00033148093064449006, + "loss": 3.3254497051239014, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.028036782518029213, + "learning_rate": 0.0003312155612164284, + "loss": 3.3276920318603516, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.028774483129382133, + "learning_rate": 0.00033095016905559883, + "loss": 3.3458826541900635, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.030403515323996544, + "learning_rate": 0.0003306847543758227, + "loss": 3.3332793712615967, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.028468478471040726, + "learning_rate": 0.0003304193173909401, + "loss": 3.3167755603790283, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.029126476496458054, + "learning_rate": 0.00033015385831480873, + "loss": 3.299426794052124, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.030746638774871826, + "learning_rate": 0.0003298883773613043, + "loss": 3.3581809997558594, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.030241351574659348, + "learning_rate": 0.0003296228747443197, + "loss": 3.3415942192077637, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.029754415154457092, + "learning_rate": 0.0003293573506777659, + "loss": 3.3525662422180176, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.026896746829152107, + "learning_rate": 0.00032909180537557076, + "loss": 3.321808099746704, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.026361066848039627, + "learning_rate": 0.00032882623905167917, + "loss": 3.329258441925049, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.027747981250286102, + "learning_rate": 0.00032856065192005335, + "loss": 3.3774523735046387, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.026270100846886635, + "learning_rate": 0.00032829504419467194, + "loss": 3.3419976234436035, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.02731695957481861, + "learning_rate": 0.0003280294160895303, + "loss": 3.3068814277648926, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.026583299040794373, + "learning_rate": 0.0003277637678186402, + "loss": 3.303804874420166, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.026615949347615242, + "learning_rate": 0.00032749809959602973, + "loss": 3.3812599182128906, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.027221621945500374, + "learning_rate": 0.0003272324116357428, + "loss": 3.327211618423462, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.025025667622685432, + "learning_rate": 0.0003269667041518395, + "loss": 3.339163303375244, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.02593836560845375, + "learning_rate": 0.0003267009773583956, + "loss": 3.3517372608184814, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.023697076365351677, + "learning_rate": 0.00032643523146950234, + "loss": 3.379819869995117, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.02425902709364891, + "learning_rate": 0.0003261694666992664, + "loss": 3.390653610229492, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.025133250281214714, + "learning_rate": 0.0003259036832618096, + "loss": 3.364495277404785, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.025542905554175377, + "learning_rate": 0.0003256378813712688, + "loss": 3.3584814071655273, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.026041125878691673, + "learning_rate": 0.000325372061241796, + "loss": 3.36777400970459, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.02391037344932556, + "learning_rate": 0.00032510622308755746, + "loss": 3.3809385299682617, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.02379394695162773, + "learning_rate": 0.00032484036712273424, + "loss": 3.354257583618164, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.0234722550958395, + "learning_rate": 0.0003245744935615219, + "loss": 3.3736205101013184, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.024123650044202805, + "learning_rate": 0.0003243086026181296, + "loss": 3.3589959144592285, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.024013696238398552, + "learning_rate": 0.00032404269450678116, + "loss": 3.353874444961548, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.022617843002080917, + "learning_rate": 0.00032377676944171375, + "loss": 3.364056348800659, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.022730547934770584, + "learning_rate": 0.00032351082763717857, + "loss": 3.378067970275879, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.022257911041378975, + "learning_rate": 0.0003232448693074399, + "loss": 3.379342555999756, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.023247094824910164, + "learning_rate": 0.00032297889466677575, + "loss": 3.3537116050720215, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.022764021530747414, + "learning_rate": 0.000322712903929477, + "loss": 3.3969063758850098, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.02260403335094452, + "learning_rate": 0.0003224468973098477, + "loss": 3.3727574348449707, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.021772351115942, + "learning_rate": 0.0003221808750222044, + "loss": 3.335434675216675, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.02238459885120392, + "learning_rate": 0.0003219148372808766, + "loss": 3.3440277576446533, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.0230540968477726, + "learning_rate": 0.00032164878430020606, + "loss": 3.3997583389282227, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.0220821350812912, + "learning_rate": 0.00032138271629454684, + "loss": 3.377815008163452, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.021497417241334915, + "learning_rate": 0.00032111663347826505, + "loss": 3.3696370124816895, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.02163885347545147, + "learning_rate": 0.00032085053606573896, + "loss": 3.3489794731140137, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.022394269704818726, + "learning_rate": 0.0003205844242713584, + "loss": 3.3792812824249268, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.022390708327293396, + "learning_rate": 0.0003203182983095248, + "loss": 3.3733272552490234, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.023712158203125, + "learning_rate": 0.0003200521583946511, + "loss": 3.3716251850128174, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.02398860827088356, + "learning_rate": 0.0003197860047411613, + "loss": 3.381779670715332, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.02309161052107811, + "learning_rate": 0.00031951983756349066, + "loss": 3.3923277854919434, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.021400388330221176, + "learning_rate": 0.00031925365707608536, + "loss": 3.407397747039795, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.022534485906362534, + "learning_rate": 0.00031898746349340204, + "loss": 3.3740391731262207, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.023533202707767487, + "learning_rate": 0.0003187212570299082, + "loss": 3.3877944946289062, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.023398835211992264, + "learning_rate": 0.0003184550379000815, + "loss": 3.3936471939086914, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.020927201956510544, + "learning_rate": 0.00031818880631841005, + "loss": 3.339545249938965, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.021453995257616043, + "learning_rate": 0.0003179225624993915, + "loss": 3.3647522926330566, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.022847114130854607, + "learning_rate": 0.0003176563066575341, + "loss": 3.405144691467285, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.022766996175050735, + "learning_rate": 0.0003173900390073549, + "loss": 3.392810344696045, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.022103948518633842, + "learning_rate": 0.0003171237597633813, + "loss": 3.3973004817962646, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.022282181307673454, + "learning_rate": 0.0003168574691401496, + "loss": 3.409998893737793, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.025435522198677063, + "learning_rate": 0.0003165911673522053, + "loss": 3.423318386077881, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.078423500061035, + "eval_runtime": 57.3189, + "eval_samples_per_second": 42.604, + "eval_steps_per_second": 1.343, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.02624760940670967, + "learning_rate": 0.00031632485461410294, + "loss": 3.261288642883301, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.03552987053990364, + "learning_rate": 0.00031605853114040583, + "loss": 3.2682106494903564, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.040380869060754776, + "learning_rate": 0.00031579219714568613, + "loss": 3.2813782691955566, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.03861293941736221, + "learning_rate": 0.0003155258528445242, + "loss": 3.258193016052246, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.04091731086373329, + "learning_rate": 0.0003152594984515089, + "loss": 3.3052220344543457, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.043269723653793335, + "learning_rate": 0.00031499313418123704, + "loss": 3.2820217609405518, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.04580320790410042, + "learning_rate": 0.0003147267602483138, + "loss": 3.312005043029785, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.047336459159851074, + "learning_rate": 0.00031446037686735153, + "loss": 3.3217713832855225, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.04587360844016075, + "learning_rate": 0.0003141939842529706, + "loss": 3.2869949340820312, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.04002532362937927, + "learning_rate": 0.0003139275826197987, + "loss": 3.2879223823547363, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.0368693508207798, + "learning_rate": 0.00031366117218247094, + "loss": 3.2983577251434326, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.037317775189876556, + "learning_rate": 0.00031339475315562916, + "loss": 3.3062524795532227, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.03655700013041496, + "learning_rate": 0.00031312832575392267, + "loss": 3.31072735786438, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.033466216176748276, + "learning_rate": 0.0003128618901920071, + "loss": 3.2954514026641846, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.030785970389842987, + "learning_rate": 0.0003125954466845446, + "loss": 3.27952241897583, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.029669342562556267, + "learning_rate": 0.00031232899544620404, + "loss": 3.2613229751586914, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.033505845814943314, + "learning_rate": 0.00031206253669166045, + "loss": 3.2987258434295654, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.032395437359809875, + "learning_rate": 0.0003117960706355947, + "loss": 3.296708583831787, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.031322430819272995, + "learning_rate": 0.0003115295974926936, + "loss": 3.3160252571105957, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.029345186427235603, + "learning_rate": 0.00031126311747765005, + "loss": 3.307143211364746, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.026715675368905067, + "learning_rate": 0.00031099663080516186, + "loss": 3.286012649536133, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.028592750430107117, + "learning_rate": 0.0003107301376899327, + "loss": 3.2528645992279053, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.028861764818429947, + "learning_rate": 0.00031046363834667123, + "loss": 3.2795748710632324, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.02609528973698616, + "learning_rate": 0.00031019713299009123, + "loss": 3.326317071914673, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.027718454599380493, + "learning_rate": 0.000309930621834911, + "loss": 3.268110513687134, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.026382867246866226, + "learning_rate": 0.0003096641050958541, + "loss": 3.294499397277832, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.027083583176136017, + "learning_rate": 0.00030939758298764795, + "loss": 3.3202130794525146, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.02516467310488224, + "learning_rate": 0.00030913105572502483, + "loss": 3.3485021591186523, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.024148069322109222, + "learning_rate": 0.00030886452352272095, + "loss": 3.3099541664123535, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.024552857503294945, + "learning_rate": 0.00030859798659547624, + "loss": 3.2885942459106445, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.02551114559173584, + "learning_rate": 0.0003083314451580349, + "loss": 3.3232040405273438, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.025423647835850716, + "learning_rate": 0.0003080648994251445, + "loss": 3.3097832202911377, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.023211417719721794, + "learning_rate": 0.0003077983496115561, + "loss": 3.3075966835021973, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.024623898789286613, + "learning_rate": 0.00030753179593202406, + "loss": 3.2917377948760986, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.02677663415670395, + "learning_rate": 0.0003072652386013059, + "loss": 3.356265068054199, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.02673642337322235, + "learning_rate": 0.00030699867783416207, + "loss": 3.314903736114502, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.024466225877404213, + "learning_rate": 0.0003067321138453557, + "loss": 3.30912446975708, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.025370748713612556, + "learning_rate": 0.0003064655468496527, + "loss": 3.363640069961548, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.026536114513874054, + "learning_rate": 0.0003061989770618211, + "loss": 3.304851531982422, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.026536893099546432, + "learning_rate": 0.00030593240469663154, + "loss": 3.342902898788452, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.02645307406783104, + "learning_rate": 0.0003056658299688563, + "loss": 3.3164432048797607, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.02523060515522957, + "learning_rate": 0.00030539925309327017, + "loss": 3.34376859664917, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.02590431459248066, + "learning_rate": 0.0003051326742846491, + "loss": 3.3479061126708984, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.0230655986815691, + "learning_rate": 0.000304866093757771, + "loss": 3.2890982627868652, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.02403893694281578, + "learning_rate": 0.00030459951172741474, + "loss": 3.345456600189209, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.02238907851278782, + "learning_rate": 0.0003043329284083609, + "loss": 3.3471767902374268, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.021377569064497948, + "learning_rate": 0.00030406634401539053, + "loss": 3.319232225418091, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.02272353693842888, + "learning_rate": 0.00030379975876328614, + "loss": 3.316892623901367, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.02413824200630188, + "learning_rate": 0.00030353317286683053, + "loss": 3.3450968265533447, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.02334682270884514, + "learning_rate": 0.0003032665865408073, + "loss": 3.3640642166137695, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.021966520696878433, + "learning_rate": 0.000303, + "loss": 3.358555316925049, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.023984894156455994, + "learning_rate": 0.0003027334134591926, + "loss": 3.358398914337158, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.023520268499851227, + "learning_rate": 0.00030246682713316945, + "loss": 3.3224587440490723, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.0231285709887743, + "learning_rate": 0.0003022002412367138, + "loss": 3.3671562671661377, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.022818170487880707, + "learning_rate": 0.00030193365598460946, + "loss": 3.361027717590332, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.02280694991350174, + "learning_rate": 0.0003016670715916391, + "loss": 3.2923383712768555, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.021362345665693283, + "learning_rate": 0.00030140048827258524, + "loss": 3.3346734046936035, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.02372218668460846, + "learning_rate": 0.0003011339062422289, + "loss": 3.354748249053955, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.022949764505028725, + "learning_rate": 0.00030086732571535076, + "loss": 3.3857789039611816, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.021954255178570747, + "learning_rate": 0.0003006007469067297, + "loss": 3.343330144882202, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.022438595071434975, + "learning_rate": 0.0003003341700311436, + "loss": 3.328618049621582, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.021969925612211227, + "learning_rate": 0.0003000675953033685, + "loss": 3.355971574783325, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.022706981748342514, + "learning_rate": 0.0002998010229381789, + "loss": 3.3529577255249023, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.024083172902464867, + "learning_rate": 0.0002995344531503474, + "loss": 3.348170757293701, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.023259354755282402, + "learning_rate": 0.0002992678861546442, + "loss": 3.3474369049072266, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.02218157798051834, + "learning_rate": 0.00029900132216583786, + "loss": 3.369053840637207, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.02305331453680992, + "learning_rate": 0.0002987347613986941, + "loss": 3.3622641563415527, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.022796178236603737, + "learning_rate": 0.0002984682040679759, + "loss": 3.351510524749756, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.021299205720424652, + "learning_rate": 0.00029820165038844395, + "loss": 3.339406728744507, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.022786924615502357, + "learning_rate": 0.00029793510057485556, + "loss": 3.3182196617126465, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.021729128435254097, + "learning_rate": 0.00029766855484196513, + "loss": 3.383904218673706, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.024604283273220062, + "learning_rate": 0.0002974020134045238, + "loss": 3.2912864685058594, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.0894880294799805, + "eval_runtime": 58.3967, + "eval_samples_per_second": 41.817, + "eval_steps_per_second": 1.319, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.02673362009227276, + "learning_rate": 0.00029713547647727904, + "loss": 3.270986795425415, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.04036599025130272, + "learning_rate": 0.00029686894427497515, + "loss": 3.2358999252319336, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.047250088304281235, + "learning_rate": 0.0002966024170123519, + "loss": 3.2706189155578613, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.04642819985747337, + "learning_rate": 0.0002963358949041459, + "loss": 3.236340284347534, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.039405327290296555, + "learning_rate": 0.00029606937816508897, + "loss": 3.203136682510376, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.038748081773519516, + "learning_rate": 0.00029580286700990887, + "loss": 3.2636070251464844, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.039233580231666565, + "learning_rate": 0.00029553636165332876, + "loss": 3.2751896381378174, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.03429320082068443, + "learning_rate": 0.0002952698623100673, + "loss": 3.2765679359436035, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.03267038241028786, + "learning_rate": 0.0002950033691948382, + "loss": 3.2387681007385254, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.03349406272172928, + "learning_rate": 0.00029473688252234994, + "loss": 3.282162666320801, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.032002877444028854, + "learning_rate": 0.0002944704025073064, + "loss": 3.246623992919922, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.031509071588516235, + "learning_rate": 0.00029420392936440525, + "loss": 3.2374932765960693, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.02958357334136963, + "learning_rate": 0.00029393746330833953, + "loss": 3.24362850189209, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.03248635679483414, + "learning_rate": 0.0002936710045537959, + "loss": 3.237210273742676, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.03230655938386917, + "learning_rate": 0.0002934045533154554, + "loss": 3.251129150390625, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.029688173905014992, + "learning_rate": 0.00029313810980799296, + "loss": 3.2141575813293457, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.027773940935730934, + "learning_rate": 0.00029287167424607726, + "loss": 3.264249086380005, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.02687031403183937, + "learning_rate": 0.00029260524684437077, + "loss": 3.2513632774353027, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.02503623627126217, + "learning_rate": 0.00029233882781752905, + "loss": 3.217961072921753, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.027071228250861168, + "learning_rate": 0.00029207241738020136, + "loss": 3.2784645557403564, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.02600773423910141, + "learning_rate": 0.0002918060157470294, + "loss": 3.2641382217407227, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.026998162269592285, + "learning_rate": 0.00029153962313264856, + "loss": 3.2646782398223877, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.025579402223229408, + "learning_rate": 0.0002912732397516862, + "loss": 3.2700581550598145, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.02645333670079708, + "learning_rate": 0.00029100686581876283, + "loss": 3.2545266151428223, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.02717350795865059, + "learning_rate": 0.0002907405015484911, + "loss": 3.2795705795288086, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.02675914391875267, + "learning_rate": 0.0002904741471554758, + "loss": 3.2936904430389404, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.02751768007874489, + "learning_rate": 0.0002902078028543139, + "loss": 3.2329440116882324, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.025446882471442223, + "learning_rate": 0.00028994146885959416, + "loss": 3.274294376373291, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.02365768514573574, + "learning_rate": 0.00028967514538589715, + "loss": 3.2585229873657227, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.0266884732991457, + "learning_rate": 0.0002894088326477947, + "loss": 3.2872962951660156, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.02572469227015972, + "learning_rate": 0.0002891425308598503, + "loss": 3.253828287124634, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.02540273219347, + "learning_rate": 0.00028887624023661866, + "loss": 3.303926944732666, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.024694625288248062, + "learning_rate": 0.00028860996099264505, + "loss": 3.264145851135254, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.027233170345425606, + "learning_rate": 0.000288343693342466, + "loss": 3.286924362182617, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02635285072028637, + "learning_rate": 0.00028807743750060847, + "loss": 3.2930331230163574, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.025157805532217026, + "learning_rate": 0.00028781119368159004, + "loss": 3.2906906604766846, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.024729596450924873, + "learning_rate": 0.00028754496209991843, + "loss": 3.277128219604492, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.023809000849723816, + "learning_rate": 0.0002872787429700917, + "loss": 3.2957217693328857, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02474421076476574, + "learning_rate": 0.0002870125365065979, + "loss": 3.27791690826416, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.02439294196665287, + "learning_rate": 0.00028674634292391463, + "loss": 3.2912981510162354, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.023601334542036057, + "learning_rate": 0.0002864801624365093, + "loss": 3.317446231842041, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.0253264419734478, + "learning_rate": 0.0002862139952588387, + "loss": 3.2742302417755127, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.02389134280383587, + "learning_rate": 0.000285947841605349, + "loss": 3.281032085418701, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.02451813407242298, + "learning_rate": 0.00028568170169047523, + "loss": 3.3339295387268066, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.024324845522642136, + "learning_rate": 0.00028541557572864154, + "loss": 3.290402889251709, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.024939008057117462, + "learning_rate": 0.000285149463934261, + "loss": 3.2933621406555176, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.024450505152344704, + "learning_rate": 0.0002848833665217349, + "loss": 3.3055014610290527, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.02262270078063011, + "learning_rate": 0.0002846172837054532, + "loss": 3.2964205741882324, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.02311946265399456, + "learning_rate": 0.0002843512156997939, + "loss": 3.3102402687072754, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.024362705647945404, + "learning_rate": 0.0002840851627191234, + "loss": 3.3020925521850586, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.024785032495856285, + "learning_rate": 0.0002838191249777955, + "loss": 3.287290096282959, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.02358846180140972, + "learning_rate": 0.00028355310269015225, + "loss": 3.334261894226074, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.024330584332346916, + "learning_rate": 0.00028328709607052297, + "loss": 3.2799081802368164, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.023546984419226646, + "learning_rate": 0.0002830211053332242, + "loss": 3.3228847980499268, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.023060832172632217, + "learning_rate": 0.0002827551306925601, + "loss": 3.325793981552124, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.02388455718755722, + "learning_rate": 0.00028248917236282147, + "loss": 3.2892062664031982, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.023740701377391815, + "learning_rate": 0.0002822232305582863, + "loss": 3.2969369888305664, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.02458217367529869, + "learning_rate": 0.0002819573054932188, + "loss": 3.315610408782959, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.023173579946160316, + "learning_rate": 0.0002816913973818703, + "loss": 3.2561936378479004, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.024137936532497406, + "learning_rate": 0.0002814255064384781, + "loss": 3.2839202880859375, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.022436272352933884, + "learning_rate": 0.00028115963287726563, + "loss": 3.3057351112365723, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.023997878655791283, + "learning_rate": 0.0002808937769124425, + "loss": 3.3583154678344727, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.024119896814227104, + "learning_rate": 0.000280627938758204, + "loss": 3.322892427444458, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.022283365949988365, + "learning_rate": 0.00028036211862873117, + "loss": 3.3145735263824463, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.023433363065123558, + "learning_rate": 0.0002800963167381904, + "loss": 3.3564529418945312, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.02224746160209179, + "learning_rate": 0.00027983053330073355, + "loss": 3.2985916137695312, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.023786570876836777, + "learning_rate": 0.00027956476853049765, + "loss": 3.302539587020874, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.025200961157679558, + "learning_rate": 0.00027929902264160435, + "loss": 3.309976577758789, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.023849809542298317, + "learning_rate": 0.00027903329584816046, + "loss": 3.343660354614258, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.023311031982302666, + "learning_rate": 0.0002787675883642572, + "loss": 3.2978909015655518, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.023174891248345375, + "learning_rate": 0.00027850190040397036, + "loss": 3.326188802719116, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.02681034244596958, + "learning_rate": 0.00027823623218135976, + "loss": 3.3422017097473145, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.109009265899658, + "eval_runtime": 58.0506, + "eval_samples_per_second": 42.067, + "eval_steps_per_second": 1.326, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.028057295829057693, + "learning_rate": 0.0002779705839104697, + "loss": 3.1825037002563477, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.034786805510520935, + "learning_rate": 0.0002777049558053281, + "loss": 3.200389862060547, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.03751441463828087, + "learning_rate": 0.0002774393480799466, + "loss": 3.2047512531280518, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.035590436309576035, + "learning_rate": 0.00027717376094832076, + "loss": 3.1735124588012695, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.03426426649093628, + "learning_rate": 0.0002769081946244293, + "loss": 3.199483871459961, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.037030890583992004, + "learning_rate": 0.00027664264932223407, + "loss": 3.213322639465332, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.032468877732753754, + "learning_rate": 0.00027637712525568024, + "loss": 3.179354190826416, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.027202758938074112, + "learning_rate": 0.0002761116226386958, + "loss": 3.216036319732666, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.031184563413262367, + "learning_rate": 0.00027584614168519125, + "loss": 3.208205223083496, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.029414812102913857, + "learning_rate": 0.00027558068260905976, + "loss": 3.180370569229126, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.02679901383817196, + "learning_rate": 0.00027531524562417734, + "loss": 3.210325002670288, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.02863631211221218, + "learning_rate": 0.00027504983094440126, + "loss": 3.1601076126098633, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.029135366901755333, + "learning_rate": 0.0002747844387835716, + "loss": 3.2262887954711914, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.030535712838172913, + "learning_rate": 0.0002745190693555099, + "loss": 3.23160719871521, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.030563877895474434, + "learning_rate": 0.0002742537228740197, + "loss": 3.255211353302002, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.02906505949795246, + "learning_rate": 0.00027398839955288575, + "loss": 3.2006797790527344, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.030871668830513954, + "learning_rate": 0.00027372309960587424, + "loss": 3.2442867755889893, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.02913421392440796, + "learning_rate": 0.0002734578232467327, + "loss": 3.218442440032959, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.030087953433394432, + "learning_rate": 0.0002731925706891891, + "loss": 3.2164535522460938, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.034147944301366806, + "learning_rate": 0.0002729273421469532, + "loss": 3.218660831451416, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.03593966364860535, + "learning_rate": 0.00027266213783371434, + "loss": 3.2240614891052246, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.032090652734041214, + "learning_rate": 0.00027239695796314327, + "loss": 3.2309441566467285, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.02776424027979374, + "learning_rate": 0.0002721318027488902, + "loss": 3.2702016830444336, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.027174271643161774, + "learning_rate": 0.00027186667240458617, + "loss": 3.19810152053833, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.028479408472776413, + "learning_rate": 0.0002716015671438419, + "loss": 3.2178421020507812, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.026820644736289978, + "learning_rate": 0.0002713364871802478, + "loss": 3.2634482383728027, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.026427917182445526, + "learning_rate": 0.0002710714327273741, + "loss": 3.2452914714813232, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.02679610252380371, + "learning_rate": 0.0002708064039987704, + "loss": 3.243825912475586, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.026115626096725464, + "learning_rate": 0.00027054140120796546, + "loss": 3.218787670135498, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.026686571538448334, + "learning_rate": 0.0002702764245684672, + "loss": 3.2785303592681885, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.02685859426856041, + "learning_rate": 0.0002700114742937627, + "loss": 3.2572226524353027, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.026968171820044518, + "learning_rate": 0.00026974655059731754, + "loss": 3.2625553607940674, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.025397246703505516, + "learning_rate": 0.0002694816536925759, + "loss": 3.262303352355957, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.025660471990704536, + "learning_rate": 0.00026921678379296057, + "loss": 3.269401788711548, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.02713891491293907, + "learning_rate": 0.0002689519411118722, + "loss": 3.2416529655456543, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.027591092512011528, + "learning_rate": 0.00026868712586269, + "loss": 3.2814316749572754, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.024462848901748657, + "learning_rate": 0.00026842233825877064, + "loss": 3.249725103378296, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.024868478998541832, + "learning_rate": 0.0002681575785134487, + "loss": 3.231574535369873, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.027015401050448418, + "learning_rate": 0.0002678928468400366, + "loss": 3.282627582550049, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.02678140625357628, + "learning_rate": 0.00026762814345182367, + "loss": 3.2208900451660156, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.026830527931451797, + "learning_rate": 0.0002673634685620767, + "loss": 3.260575294494629, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.02430890128016472, + "learning_rate": 0.0002670988223840393, + "loss": 3.2522997856140137, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.027047613635659218, + "learning_rate": 0.0002668342051309324, + "loss": 3.2721078395843506, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.026861369609832764, + "learning_rate": 0.00026656961701595306, + "loss": 3.2235946655273438, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.026650549843907356, + "learning_rate": 0.0002663050582522753, + "loss": 3.302743434906006, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.02551819011569023, + "learning_rate": 0.00026604052905304946, + "loss": 3.265763282775879, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.024843450635671616, + "learning_rate": 0.00026577602963140156, + "loss": 3.272397518157959, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.024423206225037575, + "learning_rate": 0.0002655115602004345, + "loss": 3.253495454788208, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.024180661886930466, + "learning_rate": 0.00026524712097322604, + "loss": 3.260438919067383, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.023554982617497444, + "learning_rate": 0.0002649827121628305, + "loss": 3.2342851161956787, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.023470086976885796, + "learning_rate": 0.00026471833398227694, + "loss": 3.251622438430786, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.024405231699347496, + "learning_rate": 0.0002644539866445703, + "loss": 3.2588815689086914, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.024147814139723778, + "learning_rate": 0.00026418967036269045, + "loss": 3.2723565101623535, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.023217910900712013, + "learning_rate": 0.0002639253853495921, + "loss": 3.2555792331695557, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.024921158328652382, + "learning_rate": 0.00026366113181820506, + "loss": 3.26884126663208, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.02455919235944748, + "learning_rate": 0.00026339690998143346, + "loss": 3.2674927711486816, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.024725932627916336, + "learning_rate": 0.00026313272005215637, + "loss": 3.2511110305786133, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.025769414380192757, + "learning_rate": 0.00026286856224322645, + "loss": 3.290651321411133, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.025928806513547897, + "learning_rate": 0.00026260443676747114, + "loss": 3.302396297454834, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.02499072253704071, + "learning_rate": 0.00026234034383769157, + "loss": 3.2813124656677246, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.025431277230381966, + "learning_rate": 0.0002620762836666625, + "loss": 3.234684705734253, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.026871860027313232, + "learning_rate": 0.00026181225646713257, + "loss": 3.2800331115722656, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.02578776702284813, + "learning_rate": 0.0002615482624518236, + "loss": 3.3117146492004395, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.026318378746509552, + "learning_rate": 0.0002612843018334309, + "loss": 3.257615089416504, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.025542527437210083, + "learning_rate": 0.00026102037482462256, + "loss": 3.2969868183135986, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.025723429396748543, + "learning_rate": 0.0002607564816380403, + "loss": 3.290259838104248, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.024588774889707565, + "learning_rate": 0.0002604926224862975, + "loss": 3.284090280532837, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.024619122967123985, + "learning_rate": 0.0002602287975819811, + "loss": 3.284879207611084, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.024441618472337723, + "learning_rate": 0.00025996500713765, + "loss": 3.283104658126831, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.02402760274708271, + "learning_rate": 0.0002597012513658352, + "loss": 3.3013617992401123, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.025028521195054054, + "learning_rate": 0.00025943753047904026, + "loss": 3.2892568111419678, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.029445001855492592, + "learning_rate": 0.0002591738446897401, + "loss": 3.2474446296691895, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.128176689147949, + "eval_runtime": 58.8588, + "eval_samples_per_second": 41.489, + "eval_steps_per_second": 1.308, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.0298826452344656, + "learning_rate": 0.00025891019421038177, + "loss": 3.138103485107422, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.0387548953294754, + "learning_rate": 0.00025864657925338345, + "loss": 3.161492347717285, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.03816070780158043, + "learning_rate": 0.00025838300003113514, + "loss": 3.1739673614501953, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.03409738838672638, + "learning_rate": 0.0002581194567559979, + "loss": 3.182478189468384, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.033135563135147095, + "learning_rate": 0.00025785594964030365, + "loss": 3.1720199584960938, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.03203624114394188, + "learning_rate": 0.0002575924788963554, + "loss": 3.1699862480163574, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.031237754970788956, + "learning_rate": 0.0002573290447364268, + "loss": 3.179363250732422, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.03342146426439285, + "learning_rate": 0.0002570656473727619, + "loss": 3.1822280883789062, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.03561089560389519, + "learning_rate": 0.0002568022870175752, + "loss": 3.1767094135284424, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.034710463136434555, + "learning_rate": 0.00025653896388305134, + "loss": 3.218817949295044, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.030870331451296806, + "learning_rate": 0.0002562756781813453, + "loss": 3.168602705001831, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.028004391118884087, + "learning_rate": 0.00025601243012458126, + "loss": 3.224155902862549, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.030953701585531235, + "learning_rate": 0.00025574921992485366, + "loss": 3.1789557933807373, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.030289432033896446, + "learning_rate": 0.000255486047794226, + "loss": 3.1890387535095215, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.030684132128953934, + "learning_rate": 0.0002552229139447316, + "loss": 3.195427417755127, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.03022196516394615, + "learning_rate": 0.00025495981858837235, + "loss": 3.1984965801239014, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.026935536414384842, + "learning_rate": 0.0002546967619371195, + "loss": 3.2014272212982178, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.02702922560274601, + "learning_rate": 0.0002544337442029133, + "loss": 3.205538034439087, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.026682600378990173, + "learning_rate": 0.00025417076559766194, + "loss": 3.1954591274261475, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.0273384228348732, + "learning_rate": 0.000253907826333243, + "loss": 3.1782338619232178, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.028587577864527702, + "learning_rate": 0.00025364492662150145, + "loss": 3.199565887451172, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.028848132118582726, + "learning_rate": 0.0002533820666742514, + "loss": 3.181786298751831, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.027535341680049896, + "learning_rate": 0.00025311924670327364, + "loss": 3.2013628482818604, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.02756025455892086, + "learning_rate": 0.00025285646692031804, + "loss": 3.1749587059020996, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.027358995750546455, + "learning_rate": 0.0002525937275371013, + "loss": 3.161881923675537, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.025918981060385704, + "learning_rate": 0.00025233102876530785, + "loss": 3.225224256515503, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.026479976251721382, + "learning_rate": 0.0002520683708165894, + "loss": 3.2129316329956055, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.02685079537332058, + "learning_rate": 0.0002518057539025646, + "loss": 3.1651554107666016, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.027378706261515617, + "learning_rate": 0.00025154317823481917, + "loss": 3.193108558654785, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.027240974828600883, + "learning_rate": 0.0002512806440249056, + "loss": 3.190605640411377, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.028065402060747147, + "learning_rate": 0.00025101815148434296, + "loss": 3.2068259716033936, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.028329014778137207, + "learning_rate": 0.00025075570082461684, + "loss": 3.2316060066223145, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.028920065611600876, + "learning_rate": 0.0002504932922571789, + "loss": 3.1933932304382324, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.028299354016780853, + "learning_rate": 0.00025023092599344706, + "loss": 3.192711353302002, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.025996020063757896, + "learning_rate": 0.0002499686022448049, + "loss": 3.243171453475952, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.0264595914632082, + "learning_rate": 0.00024970632122260216, + "loss": 3.199758529663086, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.02704528719186783, + "learning_rate": 0.0002494440831381538, + "loss": 3.2390713691711426, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.026571933180093765, + "learning_rate": 0.00024918188820274016, + "loss": 3.219440460205078, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.025562983006238937, + "learning_rate": 0.0002489197366276073, + "loss": 3.2538228034973145, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.024683184921741486, + "learning_rate": 0.0002486576286239657, + "loss": 3.2372591495513916, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.025656528770923615, + "learning_rate": 0.0002483955644029913, + "loss": 3.250558853149414, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.024807540699839592, + "learning_rate": 0.0002481335441758243, + "loss": 3.1856563091278076, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.026149194687604904, + "learning_rate": 0.0002478715681535699, + "loss": 3.172186851501465, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.02617705799639225, + "learning_rate": 0.00024760963654729717, + "loss": 3.250445604324341, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.026542512699961662, + "learning_rate": 0.0002473477495680397, + "loss": 3.220273494720459, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.024963509291410446, + "learning_rate": 0.00024708590742679546, + "loss": 3.2330946922302246, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.027144264429807663, + "learning_rate": 0.00024682411033452546, + "loss": 3.1860971450805664, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.02665683813393116, + "learning_rate": 0.0002465623585021553, + "loss": 3.257209062576294, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.024872658774256706, + "learning_rate": 0.0002463006521405733, + "loss": 3.2266182899475098, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.025235185399651527, + "learning_rate": 0.0002460389914606319, + "loss": 3.2134504318237305, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.025029495358467102, + "learning_rate": 0.00024577737667314615, + "loss": 3.2150087356567383, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.025630896911025047, + "learning_rate": 0.00024551580798889446, + "loss": 3.282879114151001, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.02722341939806938, + "learning_rate": 0.00024525428561861804, + "loss": 3.2366206645965576, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.02522372268140316, + "learning_rate": 0.0002449928097730207, + "loss": 3.2840523719787598, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.02494790591299534, + "learning_rate": 0.00024473138066276876, + "loss": 3.2405319213867188, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.026662737131118774, + "learning_rate": 0.0002444699984984909, + "loss": 3.239692211151123, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.026082608848810196, + "learning_rate": 0.0002442086634907782, + "loss": 3.2156195640563965, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.02423974685370922, + "learning_rate": 0.00024394737585018328, + "loss": 3.1863760948181152, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.025226688012480736, + "learning_rate": 0.00024368613578722113, + "loss": 3.20874285697937, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.02428225241601467, + "learning_rate": 0.00024342494351236799, + "loss": 3.2209601402282715, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.023819707334041595, + "learning_rate": 0.0002431637992360618, + "loss": 3.2564258575439453, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.025661710649728775, + "learning_rate": 0.00024290270316870184, + "loss": 3.244133949279785, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.02423049882054329, + "learning_rate": 0.0002426416555206484, + "loss": 3.2376630306243896, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.02382473088800907, + "learning_rate": 0.000242380656502223, + "loss": 3.2951290607452393, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.02488875202834606, + "learning_rate": 0.00024211970632370756, + "loss": 3.2848310470581055, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.02575223706662655, + "learning_rate": 0.0002418588051953453, + "loss": 3.2574503421783447, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.02462848648428917, + "learning_rate": 0.0002415979533273392, + "loss": 3.2464442253112793, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.024912388995289803, + "learning_rate": 0.0002413371509298531, + "loss": 3.2459192276000977, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.025020884349942207, + "learning_rate": 0.00024107639821301078, + "loss": 3.213622808456421, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.026330886408686638, + "learning_rate": 0.00024081569538689587, + "loss": 3.2305431365966797, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.02708602324128151, + "learning_rate": 0.0002405550426615521, + "loss": 3.2487688064575195, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.030908891931176186, + "learning_rate": 0.00024029444024698248, + "loss": 3.2110159397125244, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.146673202514648, + "eval_runtime": 58.7379, + "eval_samples_per_second": 41.574, + "eval_steps_per_second": 1.311, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.02919941022992134, + "learning_rate": 0.00024003388835314981, + "loss": 3.1533267498016357, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.03379782289266586, + "learning_rate": 0.00023977338718997595, + "loss": 3.1436538696289062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.033698830753564835, + "learning_rate": 0.00023951293696734195, + "loss": 3.15297532081604, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.03077736124396324, + "learning_rate": 0.00023925253789508806, + "loss": 3.128848075866699, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.031112324446439743, + "learning_rate": 0.00023899219018301286, + "loss": 3.1461119651794434, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.03346852958202362, + "learning_rate": 0.00023873189404087405, + "loss": 3.113943099975586, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.03307225927710533, + "learning_rate": 0.00023847164967838726, + "loss": 3.147498607635498, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.030230754986405373, + "learning_rate": 0.00023821145730522692, + "loss": 3.1162924766540527, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.02982981689274311, + "learning_rate": 0.0002379513171310251, + "loss": 3.144481658935547, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.031735632568597794, + "learning_rate": 0.00023769122936537207, + "loss": 3.133862018585205, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.0312766470015049, + "learning_rate": 0.00023743119421781614, + "loss": 3.110287666320801, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.029902562499046326, + "learning_rate": 0.00023717121189786256, + "loss": 3.1384904384613037, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.0306564774364233, + "learning_rate": 0.00023691128261497463, + "loss": 3.1138200759887695, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.03084711916744709, + "learning_rate": 0.0002366514065785725, + "loss": 3.155240535736084, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.03183220326900482, + "learning_rate": 0.00023639158399803382, + "loss": 3.1593515872955322, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.031826410442590714, + "learning_rate": 0.0002361318150826927, + "loss": 3.1517653465270996, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.030141707509756088, + "learning_rate": 0.0002358721000418404, + "loss": 3.1291589736938477, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.03158571943640709, + "learning_rate": 0.00023561243908472472, + "loss": 3.178101062774658, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.03510458394885063, + "learning_rate": 0.00023535283242054954, + "loss": 3.138767719268799, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.03660859912633896, + "learning_rate": 0.00023509328025847565, + "loss": 3.1708762645721436, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.034004807472229004, + "learning_rate": 0.00023483378280761907, + "loss": 3.1404073238372803, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.03077496401965618, + "learning_rate": 0.0002345743402770527, + "loss": 3.159081220626831, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.030384650453925133, + "learning_rate": 0.0002343149528758042, + "loss": 3.1872847080230713, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.02957277186214924, + "learning_rate": 0.00023405562081285766, + "loss": 3.165973424911499, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.029529055580496788, + "learning_rate": 0.0002337963442971522, + "loss": 3.123840808868408, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.027776138857007027, + "learning_rate": 0.00023353712353758212, + "loss": 3.1705269813537598, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.028943249955773354, + "learning_rate": 0.00023327795874299704, + "loss": 3.1181654930114746, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.028178401291370392, + "learning_rate": 0.0002330188501222012, + "loss": 3.201678991317749, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.02865665778517723, + "learning_rate": 0.0002327597978839538, + "loss": 3.18479061126709, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.029173649847507477, + "learning_rate": 0.00023250080223696852, + "loss": 3.1844916343688965, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.0280170775949955, + "learning_rate": 0.00023224186338991352, + "loss": 3.187105417251587, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.029103878885507584, + "learning_rate": 0.0002319829815514112, + "loss": 3.183281660079956, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.029045486822724342, + "learning_rate": 0.00023172415693003789, + "loss": 3.1848196983337402, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.028364380821585655, + "learning_rate": 0.000231465389734324, + "loss": 3.197348117828369, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.029651399701833725, + "learning_rate": 0.00023120668017275346, + "loss": 3.1429920196533203, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.0281438659876585, + "learning_rate": 0.00023094802845376403, + "loss": 3.1916513442993164, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.029437677934765816, + "learning_rate": 0.00023068943478574662, + "loss": 3.172391176223755, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.03433044254779816, + "learning_rate": 0.00023043089937704541, + "loss": 3.2296864986419678, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.03259492665529251, + "learning_rate": 0.00023017242243595796, + "loss": 3.1962363719940186, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.0299630519002676, + "learning_rate": 0.00022991400417073427, + "loss": 3.169673442840576, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.028495769947767258, + "learning_rate": 0.00022965564478957739, + "loss": 3.1855409145355225, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.0290420800447464, + "learning_rate": 0.0002293973445006427, + "loss": 3.23667573928833, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.03340664133429527, + "learning_rate": 0.00022913910351203816, + "loss": 3.1873788833618164, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.03253626823425293, + "learning_rate": 0.00022888092203182377, + "loss": 3.204796552658081, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.028142794966697693, + "learning_rate": 0.00022862280026801182, + "loss": 3.193922281265259, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.028805622830986977, + "learning_rate": 0.00022836473842856623, + "loss": 3.1863901615142822, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.030033612623810768, + "learning_rate": 0.00022810673672140272, + "loss": 3.198482036590576, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.028154220432043076, + "learning_rate": 0.0002278487953543889, + "loss": 3.215069055557251, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.028098132461309433, + "learning_rate": 0.00022759091453534308, + "loss": 3.2197628021240234, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.028679659590125084, + "learning_rate": 0.00022733309447203553, + "loss": 3.2062196731567383, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.029104258865118027, + "learning_rate": 0.00022707533537218695, + "loss": 3.2182908058166504, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.029629778116941452, + "learning_rate": 0.00022681763744346938, + "loss": 3.2165639400482178, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.028304366394877434, + "learning_rate": 0.00022656000089350524, + "loss": 3.1783785820007324, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.026900488883256912, + "learning_rate": 0.00022630242592986766, + "loss": 3.1876602172851562, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.027047287672758102, + "learning_rate": 0.0002260449127600803, + "loss": 3.1879465579986572, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.027424108237028122, + "learning_rate": 0.0002257874615916166, + "loss": 3.2407851219177246, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.029319508001208305, + "learning_rate": 0.00022553007263190067, + "loss": 3.1980130672454834, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.027114372700452805, + "learning_rate": 0.00022527274608830574, + "loss": 3.211357593536377, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.02579840086400509, + "learning_rate": 0.0002250154821681555, + "loss": 3.2067861557006836, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.02767331339418888, + "learning_rate": 0.0002247582810787227, + "loss": 3.198587417602539, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.026645265519618988, + "learning_rate": 0.00022450114302722957, + "loss": 3.177441120147705, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.026181906461715698, + "learning_rate": 0.0002242440682208477, + "loss": 3.226815700531006, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.027218692004680634, + "learning_rate": 0.00022398705686669748, + "loss": 3.1731510162353516, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.027551919221878052, + "learning_rate": 0.0002237301091718485, + "loss": 3.2268621921539307, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.025782009586691856, + "learning_rate": 0.00022347322534331865, + "loss": 3.2052621841430664, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.024231620132923126, + "learning_rate": 0.00022321640558807488, + "loss": 3.201770544052124, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.02474607340991497, + "learning_rate": 0.00022295965011303188, + "loss": 3.1935834884643555, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.026299327611923218, + "learning_rate": 0.00022270295912505317, + "loss": 3.223275899887085, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.02439914271235466, + "learning_rate": 0.00022244633283095, + "loss": 3.2513954639434814, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.02650653012096882, + "learning_rate": 0.0002221897714374814, + "loss": 3.1914851665496826, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.024847298860549927, + "learning_rate": 0.00022193327515135445, + "loss": 3.2173104286193848, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.028475770726799965, + "learning_rate": 0.00022167684417922337, + "loss": 3.2184863090515137, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.168707847595215, + "eval_runtime": 57.4035, + "eval_samples_per_second": 42.541, + "eval_steps_per_second": 1.341, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.02837727963924408, + "learning_rate": 0.00022142047872769014, + "loss": 3.119239330291748, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.03562312573194504, + "learning_rate": 0.00022116417900330365, + "loss": 3.092813014984131, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.0344194732606411, + "learning_rate": 0.00022090794521255997, + "loss": 3.0802855491638184, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.03188958019018173, + "learning_rate": 0.00022065177756190214, + "loss": 3.1177961826324463, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.032181013375520706, + "learning_rate": 0.00022039567625771968, + "loss": 3.1343281269073486, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.031085725873708725, + "learning_rate": 0.0002201396415063489, + "loss": 3.118828296661377, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.031995322555303574, + "learning_rate": 0.00021988367351407227, + "loss": 3.110109329223633, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.02949182130396366, + "learning_rate": 0.00021962777248711872, + "loss": 3.0972743034362793, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.029299668967723846, + "learning_rate": 0.00021937193863166292, + "loss": 3.1257238388061523, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.030558617785573006, + "learning_rate": 0.00021911617215382564, + "loss": 3.1216022968292236, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.03181575983762741, + "learning_rate": 0.00021886047325967348, + "loss": 3.093721866607666, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.030934303998947144, + "learning_rate": 0.00021860484215521805, + "loss": 3.0910396575927734, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.030235569924116135, + "learning_rate": 0.00021834927904641708, + "loss": 3.133371114730835, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.03350163996219635, + "learning_rate": 0.00021809378413917296, + "loss": 3.1178882122039795, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.03468719869852066, + "learning_rate": 0.00021783835763933344, + "loss": 3.1222469806671143, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.034203238785266876, + "learning_rate": 0.0002175829997526909, + "loss": 3.12453031539917, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.03321516141295433, + "learning_rate": 0.0002173277106849826, + "loss": 3.115428924560547, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.03298434242606163, + "learning_rate": 0.0002170724906418905, + "loss": 3.127288341522217, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.029987169429659843, + "learning_rate": 0.00021681733982904046, + "loss": 3.1194515228271484, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.031271349638700485, + "learning_rate": 0.00021656225845200332, + "loss": 3.1284165382385254, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.032201699912548065, + "learning_rate": 0.0002163072467162931, + "loss": 3.1912736892700195, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.0299740731716156, + "learning_rate": 0.00021605230482736845, + "loss": 3.1233785152435303, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.029613154008984566, + "learning_rate": 0.00021579743299063122, + "loss": 3.1535067558288574, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.03303765505552292, + "learning_rate": 0.00021554263141142726, + "loss": 3.1550910472869873, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.032644059509038925, + "learning_rate": 0.00021528790029504548, + "loss": 3.1367573738098145, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.031695179641246796, + "learning_rate": 0.00021503323984671814, + "loss": 3.1350858211517334, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.030685821548104286, + "learning_rate": 0.00021477865027162067, + "loss": 3.121593952178955, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.03091946430504322, + "learning_rate": 0.0002145241317748711, + "loss": 3.113060235977173, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.0282905176281929, + "learning_rate": 0.0002142696845615306, + "loss": 3.184551239013672, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.030020160600543022, + "learning_rate": 0.0002140153088366024, + "loss": 3.1673004627227783, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.031114062294363976, + "learning_rate": 0.00021376100480503276, + "loss": 3.162588596343994, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.03007666766643524, + "learning_rate": 0.00021350677267170973, + "loss": 3.132836103439331, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.029013793915510178, + "learning_rate": 0.00021325261264146342, + "loss": 3.1552720069885254, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.02784561738371849, + "learning_rate": 0.00021299852491906616, + "loss": 3.189272165298462, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.02759280428290367, + "learning_rate": 0.00021274450970923167, + "loss": 3.141814708709717, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.02699868381023407, + "learning_rate": 0.0002124905672166156, + "loss": 3.1127257347106934, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.02782190404832363, + "learning_rate": 0.0002122366976458146, + "loss": 3.193972587585449, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.027298318222165108, + "learning_rate": 0.00021198290120136702, + "loss": 3.1393446922302246, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.027321161702275276, + "learning_rate": 0.0002117291780877519, + "loss": 3.0926666259765625, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.026399722322821617, + "learning_rate": 0.00021147552850938941, + "loss": 3.1578259468078613, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.02770969271659851, + "learning_rate": 0.0002112219526706406, + "loss": 3.1816914081573486, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.028344258666038513, + "learning_rate": 0.00021096845077580675, + "loss": 3.1860158443450928, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.028091782703995705, + "learning_rate": 0.0002107150230291299, + "loss": 3.1585092544555664, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.026564039289951324, + "learning_rate": 0.00021046166963479204, + "loss": 3.131692409515381, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.026349127292633057, + "learning_rate": 0.00021020839079691553, + "loss": 3.160431385040283, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.030081138014793396, + "learning_rate": 0.00020995518671956256, + "loss": 3.1647543907165527, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.029250919818878174, + "learning_rate": 0.00020970205760673493, + "loss": 3.1522226333618164, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.0282477755099535, + "learning_rate": 0.00020944900366237453, + "loss": 3.139249324798584, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.0284534003585577, + "learning_rate": 0.00020919602509036195, + "loss": 3.154621124267578, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.02677394263446331, + "learning_rate": 0.0002089431220945178, + "loss": 3.1375081539154053, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.026511570438742638, + "learning_rate": 0.00020869029487860126, + "loss": 3.1355676651000977, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.02766270563006401, + "learning_rate": 0.00020843754364631068, + "loss": 3.143874168395996, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.027013977989554405, + "learning_rate": 0.0002081848686012832, + "loss": 3.1602487564086914, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.026629449799656868, + "learning_rate": 0.00020793226994709446, + "loss": 3.1549456119537354, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.0272684209048748, + "learning_rate": 0.0002076797478872588, + "loss": 3.1739020347595215, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.027645915746688843, + "learning_rate": 0.00020742730262522844, + "loss": 3.172959327697754, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.025413552299141884, + "learning_rate": 0.00020717493436439433, + "loss": 3.1649041175842285, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.028406929224729538, + "learning_rate": 0.00020692264330808455, + "loss": 3.1821887493133545, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.027753977105021477, + "learning_rate": 0.00020667042965956587, + "loss": 3.163238048553467, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.02693209983408451, + "learning_rate": 0.00020641829362204195, + "loss": 3.1921019554138184, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.02671878971159458, + "learning_rate": 0.00020616623539865446, + "loss": 3.159438133239746, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.027099428698420525, + "learning_rate": 0.00020591425519248213, + "loss": 3.1806793212890625, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.025098824873566628, + "learning_rate": 0.00020566235320654069, + "loss": 3.158310651779175, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.02655479498207569, + "learning_rate": 0.00020541052964378323, + "loss": 3.1927309036254883, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.025836164131760597, + "learning_rate": 0.00020515878470709928, + "loss": 3.1578540802001953, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.02657160349190235, + "learning_rate": 0.00020490711859931545, + "loss": 3.1658592224121094, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.027223270386457443, + "learning_rate": 0.00020465553152319418, + "loss": 3.1760241985321045, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.025695348158478737, + "learning_rate": 0.00020440402368143492, + "loss": 3.202389717102051, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.025819813832640648, + "learning_rate": 0.00020415259527667299, + "loss": 3.2111103534698486, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.02621106244623661, + "learning_rate": 0.0002039012465114796, + "loss": 3.1560299396514893, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.026585502550005913, + "learning_rate": 0.00020364997758836193, + "loss": 3.176602840423584, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.03038652054965496, + "learning_rate": 0.00020339878870976287, + "loss": 3.1799943447113037, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.193845748901367, + "eval_runtime": 57.9911, + "eval_samples_per_second": 42.11, + "eval_steps_per_second": 1.328, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.03135736659169197, + "learning_rate": 0.00020314768007806077, + "loss": 3.070986270904541, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.038986921310424805, + "learning_rate": 0.00020289665189556923, + "loss": 3.0708835124969482, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.035115309059619904, + "learning_rate": 0.00020264570436453707, + "loss": 3.1070001125335693, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.032868221402168274, + "learning_rate": 0.00020239483768714843, + "loss": 3.082258701324463, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.03703184798359871, + "learning_rate": 0.0002021440520655219, + "loss": 3.057744264602661, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.03831101953983307, + "learning_rate": 0.00020189334770171098, + "loss": 3.0789475440979004, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.036301176995038986, + "learning_rate": 0.00020164272479770352, + "loss": 3.0565457344055176, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.03166409209370613, + "learning_rate": 0.00020139218355542203, + "loss": 3.0574910640716553, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.02984458953142166, + "learning_rate": 0.000201141724176723, + "loss": 3.080018997192383, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.034001126885414124, + "learning_rate": 0.0002008913468633969, + "loss": 3.0813302993774414, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.03814122825860977, + "learning_rate": 0.00020064105181716847, + "loss": 3.0579185485839844, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.03870106488466263, + "learning_rate": 0.00020039083923969564, + "loss": 3.0857529640197754, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.04005642607808113, + "learning_rate": 0.0002001407093325704, + "loss": 3.0880727767944336, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.03760755434632301, + "learning_rate": 0.0001998906622973177, + "loss": 3.0795767307281494, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.03260907158255577, + "learning_rate": 0.0001996406983353961, + "loss": 3.1006722450256348, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.029866183176636696, + "learning_rate": 0.0001993908176481969, + "loss": 3.0850343704223633, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.030988436192274094, + "learning_rate": 0.0001991410204370445, + "loss": 3.0743892192840576, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.03698134422302246, + "learning_rate": 0.00019889130690319608, + "loss": 3.056002378463745, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.03696191683411598, + "learning_rate": 0.00019864167724784117, + "loss": 3.069124698638916, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.03453022986650467, + "learning_rate": 0.00019839213167210213, + "loss": 3.1017374992370605, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.032458193600177765, + "learning_rate": 0.00019814267037703294, + "loss": 3.1166419982910156, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.02892368845641613, + "learning_rate": 0.0001978932935636205, + "loss": 3.0863287448883057, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.030625877901911736, + "learning_rate": 0.00019764400143278276, + "loss": 3.0938801765441895, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.030280394479632378, + "learning_rate": 0.00019739479418537018, + "loss": 3.086432933807373, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.029181072488427162, + "learning_rate": 0.00019714567202216453, + "loss": 3.0977859497070312, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.028039127588272095, + "learning_rate": 0.00019689663514387896, + "loss": 3.112367868423462, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.02910272404551506, + "learning_rate": 0.00019664768375115798, + "loss": 3.1253533363342285, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.03086872771382332, + "learning_rate": 0.00019639881804457723, + "loss": 3.103689670562744, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.02883610874414444, + "learning_rate": 0.00019615003822464342, + "loss": 3.111599922180176, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.028490116819739342, + "learning_rate": 0.00019590134449179371, + "loss": 3.1410303115844727, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.028714407235383987, + "learning_rate": 0.00019565273704639648, + "loss": 3.133023738861084, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.029553238302469254, + "learning_rate": 0.00019540421608874994, + "loss": 3.107003688812256, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.02997349575161934, + "learning_rate": 0.0001951557818190831, + "loss": 3.135680675506592, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.029087591916322708, + "learning_rate": 0.00019490743443755504, + "loss": 3.1103882789611816, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.029947761446237564, + "learning_rate": 0.0001946591741442546, + "loss": 3.0961623191833496, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.02950555644929409, + "learning_rate": 0.00019441100113920076, + "loss": 3.1162495613098145, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.029569562524557114, + "learning_rate": 0.00019416291562234192, + "loss": 3.096372127532959, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.031366728246212006, + "learning_rate": 0.00019391491779355608, + "loss": 3.120089530944824, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.0313473716378212, + "learning_rate": 0.00019366700785265065, + "loss": 3.100785732269287, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.028021851554512978, + "learning_rate": 0.0001934191859993622, + "loss": 3.0940558910369873, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.027477117255330086, + "learning_rate": 0.00019317145243335642, + "loss": 3.1210484504699707, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.029690749943256378, + "learning_rate": 0.0001929238073542276, + "loss": 3.1350059509277344, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.029659733176231384, + "learning_rate": 0.00019267625096149903, + "loss": 3.1220788955688477, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.028260033577680588, + "learning_rate": 0.00019242878345462227, + "loss": 3.1000189781188965, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.027554461732506752, + "learning_rate": 0.00019218140503297765, + "loss": 3.1044349670410156, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.027929967269301414, + "learning_rate": 0.00019193411589587332, + "loss": 3.1353745460510254, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.029309319332242012, + "learning_rate": 0.00019168691624254564, + "loss": 3.1578292846679688, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.029183287173509598, + "learning_rate": 0.00019143980627215915, + "loss": 3.1217150688171387, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.02707512304186821, + "learning_rate": 0.00019119278618380555, + "loss": 3.119617223739624, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.03006536327302456, + "learning_rate": 0.00019094585617650476, + "loss": 3.1481480598449707, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.03026784397661686, + "learning_rate": 0.00019069901644920366, + "loss": 3.142770767211914, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.02827967144548893, + "learning_rate": 0.00019045226720077667, + "loss": 3.1582541465759277, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.028892774134874344, + "learning_rate": 0.0001902056086300251, + "loss": 3.1530709266662598, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.03023783676326275, + "learning_rate": 0.0001899590409356773, + "loss": 3.1102521419525146, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.027951840311288834, + "learning_rate": 0.00018971256431638854, + "loss": 3.139113426208496, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.02807226963341236, + "learning_rate": 0.00018946617897074041, + "loss": 3.149428367614746, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.029721451923251152, + "learning_rate": 0.0001892198850972413, + "loss": 3.120518207550049, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.02867680788040161, + "learning_rate": 0.00018897368289432553, + "loss": 3.1109931468963623, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.02663845382630825, + "learning_rate": 0.00018872757256035403, + "loss": 3.116053581237793, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.028366081416606903, + "learning_rate": 0.00018848155429361323, + "loss": 3.1444268226623535, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.028931576758623123, + "learning_rate": 0.00018823562829231567, + "loss": 3.163870096206665, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.02664567157626152, + "learning_rate": 0.0001879897947545996, + "loss": 3.1349146366119385, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.028031354770064354, + "learning_rate": 0.00018774405387852848, + "loss": 3.1638219356536865, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.029421644285321236, + "learning_rate": 0.00018749840586209144, + "loss": 3.1601340770721436, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.027983693405985832, + "learning_rate": 0.00018725285090320245, + "loss": 3.1636788845062256, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.02719871513545513, + "learning_rate": 0.00018700738919970103, + "loss": 3.1246869564056396, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.03030397929251194, + "learning_rate": 0.00018676202094935085, + "loss": 3.1749579906463623, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.029340893030166626, + "learning_rate": 0.000186516746349841, + "loss": 3.1751790046691895, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.027624106034636497, + "learning_rate": 0.0001862715655987847, + "loss": 3.178704023361206, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.02830546908080578, + "learning_rate": 0.00018602647889371957, + "loss": 3.177117347717285, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.03168068826198578, + "learning_rate": 0.00018578148643210764, + "loss": 3.125427722930908, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.033327918499708176, + "learning_rate": 0.00018553658841133483, + "loss": 3.13071870803833, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.218388080596924, + "eval_runtime": 57.611, + "eval_samples_per_second": 42.388, + "eval_steps_per_second": 1.337, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.03336749225854874, + "learning_rate": 0.00018529178502871117, + "loss": 3.064338445663452, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.038479577749967575, + "learning_rate": 0.0001850470764814702, + "loss": 3.0478687286376953, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.032562077045440674, + "learning_rate": 0.0001848024629667691, + "loss": 3.019803762435913, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.0331728458404541, + "learning_rate": 0.00018455794468168887, + "loss": 3.0599141120910645, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.03361169621348381, + "learning_rate": 0.00018431352182323322, + "loss": 3.0315823554992676, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.032973092049360275, + "learning_rate": 0.0001840691945883293, + "loss": 3.025928258895874, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.030388766899704933, + "learning_rate": 0.00018382496317382715, + "loss": 3.070671558380127, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.030899515375494957, + "learning_rate": 0.0001835808277764997, + "loss": 3.012938976287842, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.0335412323474884, + "learning_rate": 0.00018333678859304222, + "loss": 3.0156331062316895, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.035731345415115356, + "learning_rate": 0.00018309284582007274, + "loss": 3.0578935146331787, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.03379053995013237, + "learning_rate": 0.0001828489996541318, + "loss": 3.048701286315918, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.031242679804563522, + "learning_rate": 0.00018260525029168147, + "loss": 3.0832977294921875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.030997611582279205, + "learning_rate": 0.00018236159792910648, + "loss": 3.0663583278656006, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.03397331386804581, + "learning_rate": 0.00018211804276271304, + "loss": 3.0115303993225098, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.03349050134420395, + "learning_rate": 0.00018187458498872914, + "loss": 3.02864933013916, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.031079065054655075, + "learning_rate": 0.00018163122480330433, + "loss": 3.053401470184326, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.029346197843551636, + "learning_rate": 0.00018138796240250955, + "loss": 3.0700926780700684, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.030784208327531815, + "learning_rate": 0.00018114479798233686, + "loss": 3.066525936126709, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.03368140384554863, + "learning_rate": 0.00018090173173869939, + "loss": 3.0956242084503174, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.03356792405247688, + "learning_rate": 0.00018065876386743143, + "loss": 3.0394320487976074, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.029909325763583183, + "learning_rate": 0.00018041589456428754, + "loss": 3.031525135040283, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.030711578205227852, + "learning_rate": 0.0001801731240249434, + "loss": 3.075474977493286, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.03452072665095329, + "learning_rate": 0.00017993045244499463, + "loss": 3.0228147506713867, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.03235075995326042, + "learning_rate": 0.00017968788001995742, + "loss": 3.0664501190185547, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.029671598225831985, + "learning_rate": 0.000179445406945268, + "loss": 3.055302143096924, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.033868469297885895, + "learning_rate": 0.0001792030334162825, + "loss": 3.053016185760498, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.0329158715903759, + "learning_rate": 0.00017896075962827706, + "loss": 3.1017332077026367, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.031983695924282074, + "learning_rate": 0.00017871858577644712, + "loss": 3.0706472396850586, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.02963697351515293, + "learning_rate": 0.0001784765120559078, + "loss": 3.0711426734924316, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.03162391856312752, + "learning_rate": 0.00017823453866169355, + "loss": 3.044867515563965, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.030773654580116272, + "learning_rate": 0.00017799266578875808, + "loss": 3.0963196754455566, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.031220417469739914, + "learning_rate": 0.00017775089363197371, + "loss": 3.0892245769500732, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.030928561463952065, + "learning_rate": 0.00017750922238613198, + "loss": 3.1181323528289795, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.03192073106765747, + "learning_rate": 0.00017726765224594342, + "loss": 3.066584348678589, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.03075495734810829, + "learning_rate": 0.0001770261834060363, + "loss": 3.0436511039733887, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.02895325981080532, + "learning_rate": 0.00017678481606095786, + "loss": 3.123373031616211, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.029738714918494225, + "learning_rate": 0.0001765435504051734, + "loss": 3.1049060821533203, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.031937118619680405, + "learning_rate": 0.0001763023866330664, + "loss": 3.099111318588257, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.029886579141020775, + "learning_rate": 0.00017606132493893793, + "loss": 3.0970301628112793, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.029569782316684723, + "learning_rate": 0.00017582036551700724, + "loss": 3.062499761581421, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.02924283780157566, + "learning_rate": 0.00017557950856141085, + "loss": 3.0618538856506348, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.030955800786614418, + "learning_rate": 0.00017533875426620286, + "loss": 3.0745606422424316, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.03145027533173561, + "learning_rate": 0.00017509810282535497, + "loss": 3.122938632965088, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.028926467522978783, + "learning_rate": 0.00017485755443275513, + "loss": 3.0808773040771484, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.029592076316475868, + "learning_rate": 0.0001746171092822094, + "loss": 3.1167049407958984, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.028070243075489998, + "learning_rate": 0.00017437676756743974, + "loss": 3.0781636238098145, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.028642063960433006, + "learning_rate": 0.00017413652948208524, + "loss": 3.093937397003174, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.0290757454931736, + "learning_rate": 0.00017389639521970136, + "loss": 3.1079559326171875, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.027867283672094345, + "learning_rate": 0.00017365636497376004, + "loss": 3.0885753631591797, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.02896495908498764, + "learning_rate": 0.00017341643893764927, + "loss": 3.0889246463775635, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.02777806483209133, + "learning_rate": 0.000173176617304673, + "loss": 3.084439277648926, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.02904277853667736, + "learning_rate": 0.00017293690026805124, + "loss": 3.125246047973633, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.029010627418756485, + "learning_rate": 0.00017269728802091975, + "loss": 3.111422061920166, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.028178894892334938, + "learning_rate": 0.00017245778075632975, + "loss": 3.115067481994629, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.02763799950480461, + "learning_rate": 0.00017221837866724792, + "loss": 3.1228506565093994, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.028052521869540215, + "learning_rate": 0.00017197908194655615, + "loss": 3.1150546073913574, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.029438253492116928, + "learning_rate": 0.00017173989078705175, + "loss": 3.0921788215637207, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.028195543214678764, + "learning_rate": 0.0001715008053814463, + "loss": 3.1408450603485107, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.02937602624297142, + "learning_rate": 0.00017126182592236682, + "loss": 3.111612319946289, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.028430236503481865, + "learning_rate": 0.00017102295260235472, + "loss": 3.093441963195801, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.027624569833278656, + "learning_rate": 0.0001707841856138659, + "loss": 3.1184372901916504, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.029797039926052094, + "learning_rate": 0.00017054552514927061, + "loss": 3.1074559688568115, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.028728218749165535, + "learning_rate": 0.00017030697140085334, + "loss": 3.161461114883423, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.02704412117600441, + "learning_rate": 0.0001700685245608126, + "loss": 3.13325834274292, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.028067929670214653, + "learning_rate": 0.00016983018482126047, + "loss": 3.093172550201416, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.028848597779870033, + "learning_rate": 0.00016959195237422305, + "loss": 3.1461825370788574, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.02956785447895527, + "learning_rate": 0.00016935382741164, + "loss": 3.141045570373535, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.028214773163199425, + "learning_rate": 0.00016911581012536424, + "loss": 3.136033773422241, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.029155246913433075, + "learning_rate": 0.0001688779007071622, + "loss": 3.0986382961273193, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.02968468703329563, + "learning_rate": 0.00016864009934871275, + "loss": 3.11630916595459, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.028356682509183884, + "learning_rate": 0.0001684024062416087, + "loss": 3.1134822368621826, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.03301718086004257, + "learning_rate": 0.0001681648215773546, + "loss": 3.118591785430908, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.246211528778076, + "eval_runtime": 57.5155, + "eval_samples_per_second": 42.458, + "eval_steps_per_second": 1.339, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.033637888729572296, + "learning_rate": 0.0001679273455473684, + "loss": 3.0192506313323975, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.04057888314127922, + "learning_rate": 0.0001676899783429802, + "loss": 2.977440595626831, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.03579854592680931, + "learning_rate": 0.00016745272015543242, + "loss": 2.9902219772338867, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.03131850063800812, + "learning_rate": 0.0001672155711758799, + "loss": 2.9988656044006348, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.03262636810541153, + "learning_rate": 0.00016697853159538896, + "loss": 3.037747621536255, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.03260403499007225, + "learning_rate": 0.0001667416016049383, + "loss": 2.987508773803711, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.033165134489536285, + "learning_rate": 0.0001665047813954181, + "loss": 3.0511183738708496, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.03303489834070206, + "learning_rate": 0.00016626807115763023, + "loss": 3.0083041191101074, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.03185006603598595, + "learning_rate": 0.00016603147108228754, + "loss": 2.9966228008270264, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.03135651350021362, + "learning_rate": 0.00016579498136001485, + "loss": 3.036759376525879, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.03216925635933876, + "learning_rate": 0.00016555860218134738, + "loss": 3.0304150581359863, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.03231671452522278, + "learning_rate": 0.00016532233373673164, + "loss": 3.0300474166870117, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.03263348713517189, + "learning_rate": 0.00016508617621652488, + "loss": 3.0298163890838623, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.032594650983810425, + "learning_rate": 0.00016485012981099488, + "loss": 2.98789119720459, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.03406650573015213, + "learning_rate": 0.0001646141947103201, + "loss": 3.0102760791778564, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.03379780799150467, + "learning_rate": 0.0001643783711045889, + "loss": 3.0564732551574707, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.03172587603330612, + "learning_rate": 0.0001641426591838005, + "loss": 3.0410943031311035, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.02938298135995865, + "learning_rate": 0.00016390705913786344, + "loss": 3.0517544746398926, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.03206252679228783, + "learning_rate": 0.00016367157115659646, + "loss": 3.0352776050567627, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.033386752009391785, + "learning_rate": 0.00016343619542972808, + "loss": 3.020535945892334, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.033984702080488205, + "learning_rate": 0.0001632009321468962, + "loss": 3.0552730560302734, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.0329931266605854, + "learning_rate": 0.00016296578149764833, + "loss": 3.0075573921203613, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.032101184129714966, + "learning_rate": 0.00016273074367144087, + "loss": 3.046388864517212, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.032873060554265976, + "learning_rate": 0.00016249581885763968, + "loss": 3.0340287685394287, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.030672110617160797, + "learning_rate": 0.0001622610072455194, + "loss": 3.0347838401794434, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.030014054849743843, + "learning_rate": 0.00016202630902426355, + "loss": 3.0315332412719727, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.02932834066450596, + "learning_rate": 0.00016179172438296423, + "loss": 3.084340810775757, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.029181934893131256, + "learning_rate": 0.00016155725351062205, + "loss": 3.032886028289795, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.02966221794486046, + "learning_rate": 0.00016132289659614605, + "loss": 3.032766819000244, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.029713116586208344, + "learning_rate": 0.00016108865382835318, + "loss": 3.028778076171875, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.029243769124150276, + "learning_rate": 0.00016085452539596864, + "loss": 3.0492146015167236, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.02997511625289917, + "learning_rate": 0.0001606205114876256, + "loss": 3.0772786140441895, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.02930292673408985, + "learning_rate": 0.00016038661229186472, + "loss": 3.055750608444214, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.03037521243095398, + "learning_rate": 0.00016015282799713453, + "loss": 3.033710479736328, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.029957240447402, + "learning_rate": 0.0001599191587917905, + "loss": 3.0394721031188965, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.02925589308142662, + "learning_rate": 0.0001596856048640961, + "loss": 3.0878353118896484, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.03003518097102642, + "learning_rate": 0.00015945216640222116, + "loss": 3.048628330230713, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.031545430421829224, + "learning_rate": 0.000159218843594243, + "loss": 3.0775623321533203, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.031310439109802246, + "learning_rate": 0.0001589856366281455, + "loss": 3.01645565032959, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.03074057213962078, + "learning_rate": 0.0001587525456918194, + "loss": 3.0651440620422363, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.028704499825835228, + "learning_rate": 0.0001585195709730619, + "loss": 3.0527148246765137, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.029061976820230484, + "learning_rate": 0.00015828671265957623, + "loss": 3.0854501724243164, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.03034680150449276, + "learning_rate": 0.00015805397093897255, + "loss": 3.056356906890869, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.028715113177895546, + "learning_rate": 0.00015782134599876633, + "loss": 3.0874247550964355, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.030256496742367744, + "learning_rate": 0.00015758883802637953, + "loss": 3.052424430847168, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.02949833683669567, + "learning_rate": 0.00015735644720913923, + "loss": 3.0663650035858154, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.028980428352952003, + "learning_rate": 0.00015712417373427897, + "loss": 3.0700113773345947, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.028526026755571365, + "learning_rate": 0.00015689201778893713, + "loss": 3.0856711864471436, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.029572727158665657, + "learning_rate": 0.00015665997956015743, + "loss": 3.0862035751342773, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.02955584041774273, + "learning_rate": 0.00015642805923488895, + "loss": 3.030802011489868, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.02881692908704281, + "learning_rate": 0.00015619625699998585, + "loss": 3.074392318725586, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.029564999043941498, + "learning_rate": 0.00015596457304220692, + "loss": 3.057894229888916, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.030315004289150238, + "learning_rate": 0.00015573300754821547, + "loss": 3.096233367919922, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.030666006729006767, + "learning_rate": 0.00015550156070458002, + "loss": 3.040799617767334, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.028901133686304092, + "learning_rate": 0.0001552702326977731, + "loss": 3.055051803588867, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.030359387397766113, + "learning_rate": 0.00015503902371417122, + "loss": 3.085972309112549, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.03117106668651104, + "learning_rate": 0.00015480793394005547, + "loss": 3.0786352157592773, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.029238386079669, + "learning_rate": 0.0001545769635616107, + "loss": 3.0375771522521973, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.029324093833565712, + "learning_rate": 0.0001543461127649256, + "loss": 3.1140198707580566, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.02875387854874134, + "learning_rate": 0.00015411538173599231, + "loss": 3.090881109237671, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.029023943468928337, + "learning_rate": 0.00015388477066070674, + "loss": 3.0755701065063477, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.0291463453322649, + "learning_rate": 0.00015365427972486827, + "loss": 3.0723776817321777, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.02795177884399891, + "learning_rate": 0.00015342390911417904, + "loss": 3.058993339538574, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.029499078169465065, + "learning_rate": 0.00015319365901424452, + "loss": 3.0856404304504395, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.029889609664678574, + "learning_rate": 0.0001529635296105732, + "loss": 3.0622942447662354, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.03107610158622265, + "learning_rate": 0.0001527335210885761, + "loss": 3.114957332611084, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.02967257983982563, + "learning_rate": 0.00015250363363356685, + "loss": 3.088871479034424, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.029068559408187866, + "learning_rate": 0.00015227386743076155, + "loss": 3.1103081703186035, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.029334546998143196, + "learning_rate": 0.00015204422266527911, + "loss": 3.0841622352600098, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.029259610921144485, + "learning_rate": 0.00015181469952213968, + "loss": 3.0662848949432373, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.030072513967752457, + "learning_rate": 0.0001515852981862663, + "loss": 3.0900328159332275, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.033321622759103775, + "learning_rate": 0.00015135601884248303, + "loss": 3.0871129035949707, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.271463394165039, + "eval_runtime": 57.9048, + "eval_samples_per_second": 42.173, + "eval_steps_per_second": 1.33, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.03635408729314804, + "learning_rate": 0.0001511268616755166, + "loss": 2.9576284885406494, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.03927835449576378, + "learning_rate": 0.00015089782686999445, + "loss": 2.999624729156494, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.03371119126677513, + "learning_rate": 0.00015066891461044588, + "loss": 2.971052646636963, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.033049341291189194, + "learning_rate": 0.00015044012508130138, + "loss": 2.980811595916748, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.03830729424953461, + "learning_rate": 0.00015021145846689254, + "loss": 2.9985947608947754, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.036573827266693115, + "learning_rate": 0.00014998291495145206, + "loss": 2.9892568588256836, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.031196987256407738, + "learning_rate": 0.00014975449471911294, + "loss": 2.9854254722595215, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.03376495838165283, + "learning_rate": 0.0001495261979539097, + "loss": 2.9840264320373535, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.036272983998060226, + "learning_rate": 0.0001492980248397766, + "loss": 3.0168871879577637, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.03328242897987366, + "learning_rate": 0.00014906997556054865, + "loss": 2.9815053939819336, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.03158426657319069, + "learning_rate": 0.00014884205029996113, + "loss": 2.994525909423828, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.03320910036563873, + "learning_rate": 0.00014861424924164925, + "loss": 3.008638858795166, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.03160254284739494, + "learning_rate": 0.00014838657256914832, + "loss": 2.979243040084839, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.03158991411328316, + "learning_rate": 0.0001481590204658931, + "loss": 2.938634157180786, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.031116485595703125, + "learning_rate": 0.00014793159311521833, + "loss": 2.9957656860351562, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.030757686123251915, + "learning_rate": 0.00014770429070035812, + "loss": 3.0106635093688965, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.02992827817797661, + "learning_rate": 0.00014747711340444601, + "loss": 2.996793746948242, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.0330309197306633, + "learning_rate": 0.00014725006141051434, + "loss": 2.9859161376953125, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.029488038271665573, + "learning_rate": 0.00014702313490149512, + "loss": 3.0081536769866943, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.03095114603638649, + "learning_rate": 0.00014679633406021892, + "loss": 3.006385087966919, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.03194110840559006, + "learning_rate": 0.00014656965906941494, + "loss": 3.0125250816345215, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.03306785598397255, + "learning_rate": 0.00014634311011171112, + "loss": 2.9640283584594727, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.030883541330695152, + "learning_rate": 0.00014611668736963396, + "loss": 3.0223653316497803, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.03224025294184685, + "learning_rate": 0.00014589039102560822, + "loss": 3.0274910926818848, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.03182433918118477, + "learning_rate": 0.00014566422126195656, + "loss": 3.047736406326294, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.03387221321463585, + "learning_rate": 0.00014543817826089992, + "loss": 2.9821343421936035, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.03407924994826317, + "learning_rate": 0.0001452122622045574, + "loss": 3.041207790374756, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.031546324491500854, + "learning_rate": 0.00014498647327494506, + "loss": 3.0298609733581543, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.03354770317673683, + "learning_rate": 0.00014476081165397715, + "loss": 3.0212411880493164, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.031692177057266235, + "learning_rate": 0.00014453527752346517, + "loss": 3.007350444793701, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.03218817710876465, + "learning_rate": 0.00014430987106511795, + "loss": 3.0019829273223877, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.03233771771192551, + "learning_rate": 0.00014408459246054126, + "loss": 3.0400614738464355, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.028998777270317078, + "learning_rate": 0.00014385944189123794, + "loss": 3.022221088409424, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.030422436073422432, + "learning_rate": 0.00014363441953860805, + "loss": 3.0071797370910645, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.030207227915525436, + "learning_rate": 0.00014340952558394778, + "loss": 2.9762964248657227, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.0305719543248415, + "learning_rate": 0.00014318476020845018, + "loss": 3.075176239013672, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.03173545375466347, + "learning_rate": 0.0001429601235932047, + "loss": 3.028982162475586, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.02799576334655285, + "learning_rate": 0.00014273561591919705, + "loss": 3.0242385864257812, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.03140150383114815, + "learning_rate": 0.00014251123736730884, + "loss": 3.012221336364746, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.029416466131806374, + "learning_rate": 0.00014228698811831787, + "loss": 3.0037949085235596, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.028385788202285767, + "learning_rate": 0.0001420628683528978, + "loss": 3.0350992679595947, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.031597770750522614, + "learning_rate": 0.00014183887825161779, + "loss": 3.065139055252075, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.02926146425306797, + "learning_rate": 0.0001416150179949428, + "loss": 3.014230728149414, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.02970231883227825, + "learning_rate": 0.0001413912877632326, + "loss": 3.06636381149292, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.029684338718652725, + "learning_rate": 0.00014116768773674306, + "loss": 3.024610996246338, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.02908361330628395, + "learning_rate": 0.00014094421809562438, + "loss": 3.0093719959259033, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.029891187325119972, + "learning_rate": 0.00014072087901992209, + "loss": 3.0735411643981934, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.028607673943042755, + "learning_rate": 0.00014049767068957645, + "loss": 3.044274091720581, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.029806701466441154, + "learning_rate": 0.00014027459328442242, + "loss": 3.0449328422546387, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.03161340951919556, + "learning_rate": 0.00014005164698418948, + "loss": 3.046813488006592, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.028860243037343025, + "learning_rate": 0.0001398288319685013, + "loss": 3.011828660964966, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.0285879448056221, + "learning_rate": 0.00013960614841687595, + "loss": 3.043640613555908, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.0296401996165514, + "learning_rate": 0.00013938359650872558, + "loss": 3.071589469909668, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.02807026542723179, + "learning_rate": 0.00013916117642335623, + "loss": 3.018320083618164, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.029037052765488625, + "learning_rate": 0.00013893888833996774, + "loss": 3.0471763610839844, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.029226379469037056, + "learning_rate": 0.00013871673243765362, + "loss": 3.037463426589966, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.029322804883122444, + "learning_rate": 0.00013849470889540098, + "loss": 3.0472755432128906, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.028253905475139618, + "learning_rate": 0.00013827281789208996, + "loss": 3.0265414714813232, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.028875533491373062, + "learning_rate": 0.00013805105960649423, + "loss": 3.063941717147827, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.028460513800382614, + "learning_rate": 0.00013782943421728042, + "loss": 3.0512683391571045, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.02951696142554283, + "learning_rate": 0.00013760794190300827, + "loss": 3.0528573989868164, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.030593490228056908, + "learning_rate": 0.00013738658284212973, + "loss": 3.057919979095459, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.028074532747268677, + "learning_rate": 0.00013716535721299016, + "loss": 3.0263280868530273, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.02955678664147854, + "learning_rate": 0.0001369442651938271, + "loss": 3.081322193145752, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.029724644497036934, + "learning_rate": 0.00013672330696277014, + "loss": 3.059583902359009, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.029713226482272148, + "learning_rate": 0.00013650248269784143, + "loss": 3.018678903579712, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.029758386313915253, + "learning_rate": 0.00013628179257695508, + "loss": 3.0379555225372314, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.028971998021006584, + "learning_rate": 0.00013606123677791727, + "loss": 3.041267156600952, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.02968725562095642, + "learning_rate": 0.00013584081547842565, + "loss": 3.0675506591796875, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.029876509681344032, + "learning_rate": 0.00013562052885606957, + "loss": 3.051586151123047, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.02853119559586048, + "learning_rate": 0.00013540037708833043, + "loss": 3.0335168838500977, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.03374241292476654, + "learning_rate": 0.00013518036035258012, + "loss": 3.027909278869629, + "step": 2520 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.0743646764662784e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-2520/training_args.bin b/runs/l2r90-baseline/checkpoint-2520/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2520/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-2880/chat_template.jinja b/runs/l2r90-baseline/checkpoint-2880/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-2880/config.json b/runs/l2r90-baseline/checkpoint-2880/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-2880/generation_config.json b/runs/l2r90-baseline/checkpoint-2880/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-2880/model.safetensors b/runs/l2r90-baseline/checkpoint-2880/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cbb2d4c9c0c87889c2a44418aa3e5c95e013a74e --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:516c4b11ac1bff4c61f8f90e9544c9b2e5a70b2b0a2026afcd2343104a0742b7 +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-2880/optimizer.pt b/runs/l2r90-baseline/checkpoint-2880/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..74f8275f45f0188f970a6fd417cc9709275e954d --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e5bc533ab14e34fccd687fe60d700ddf357ff476ef17cd6525caf7beaa1b701 +size 1166827898 diff --git a/runs/l2r90-baseline/checkpoint-2880/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-2880/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..0f2f3f214d81f9b4c1f52472b6c90d59a40880d9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d02dbaa2c75ede2796f9723198b9e46b03b4e1acd7596aef25899b19685b472 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2880/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-2880/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..421d49be31deb98414023cc92e42baeafe9660da --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1b031e812ec63375a07102faa7694fd9d3b4de75bec687a01240ccef400a542 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2880/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-2880/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..10dca30de14d8c99876de5aa078e8c5fe6e4805c --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d37e0894989a7ddb416a70e836d2c09c377786a724fba5c284f1b245a61a12b +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2880/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-2880/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..5b94d5903005d6524a67bfcf0219123f4e2e99df --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c9e6be7a3ea88e462b65bb71838108dcecb9f6c8bdaf41a50c1130618751a09 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-2880/scheduler.pt b/runs/l2r90-baseline/checkpoint-2880/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..5c7f6b411db3da6cc58af5b4530e84dddd5abe00 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e88932b62643ef7a711b903ac04665eab5638dccef14b4bec61d20b2791e67b +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-2880/tokenizer.json b/runs/l2r90-baseline/checkpoint-2880/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-2880/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-2880/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-2880/trainer_state.json b/runs/l2r90-baseline/checkpoint-2880/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..9c8d559ac440a4458709b238f3047f9de6460966 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/trainer_state.json @@ -0,0 +1,20498 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 40.0, + "eval_steps": 500, + "global_step": 2880, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 5.549652099609375, + "eval_runtime": 57.761, + "eval_samples_per_second": 42.278, + "eval_steps_per_second": 1.333, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.03958690166473389, + "learning_rate": 0.0005919487226162231, + "loss": 5.459286689758301, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.03299890458583832, + "learning_rate": 0.0005918869546699877, + "loss": 5.45464563369751, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.028040863573551178, + "learning_rate": 0.0005918249539727509, + "loss": 5.444796562194824, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.028365720063447952, + "learning_rate": 0.0005917627205744655, + "loss": 5.447852611541748, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.02822224237024784, + "learning_rate": 0.0005917002545252719, + "loss": 5.453883171081543, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.036075398325920105, + "learning_rate": 0.0005916375558754977, + "loss": 5.462671279907227, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04683876410126686, + "learning_rate": 0.0005915746246756581, + "loss": 5.443917274475098, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.05457588657736778, + "learning_rate": 0.0005915114609764558, + "loss": 5.444374084472656, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.04250836744904518, + "learning_rate": 0.0005914480648287804, + "loss": 5.435029983520508, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.029477844014763832, + "learning_rate": 0.0005913844362837093, + "loss": 5.421862602233887, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.03673284500837326, + "learning_rate": 0.0005913205753925066, + "loss": 5.430391311645508, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.030489861965179443, + "learning_rate": 0.0005912564822066241, + "loss": 5.389314651489258, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.04037046805024147, + "learning_rate": 0.0005911921567777005, + "loss": 5.425060272216797, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.0445217490196228, + "learning_rate": 0.0005911275991575613, + "loss": 5.437565326690674, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.03885014355182648, + "learning_rate": 0.0005910628093982198, + "loss": 5.410930633544922, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.0334964245557785, + "learning_rate": 0.0005909977875518759, + "loss": 5.412092208862305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.03499286621809006, + "learning_rate": 0.0005909325336709164, + "loss": 5.382216453552246, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.04171644523739815, + "learning_rate": 0.0005908670478079152, + "loss": 5.387385368347168, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.057558078318834305, + "learning_rate": 0.000590801330015633, + "loss": 5.437925815582275, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.053846780210733414, + "learning_rate": 0.0005907353803470177, + "loss": 5.41644811630249, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.027259036898612976, + "learning_rate": 0.0005906691988552034, + "loss": 5.409185409545898, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.03394043818116188, + "learning_rate": 0.0005906027855935115, + "loss": 5.377973556518555, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.03731574863195419, + "learning_rate": 0.0005905361406154501, + "loss": 5.382505893707275, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.04705141484737396, + "learning_rate": 0.0005904692639747137, + "loss": 5.413661003112793, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.044811323285102844, + "learning_rate": 0.0005904021557251837, + "loss": 5.417835235595703, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.04000752419233322, + "learning_rate": 0.0005903348159209277, + "loss": 5.406160354614258, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.038364164531230927, + "learning_rate": 0.0005902672446162007, + "loss": 5.357950687408447, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.030500127002596855, + "learning_rate": 0.0005901994418654432, + "loss": 5.344951629638672, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.030644729733467102, + "learning_rate": 0.0005901314077232829, + "loss": 5.3599348068237305, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.037976790219545364, + "learning_rate": 0.0005900631422445335, + "loss": 5.354084491729736, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.04106011986732483, + "learning_rate": 0.0005899946454841955, + "loss": 5.363678455352783, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.04334832355380058, + "learning_rate": 0.0005899259174974553, + "loss": 5.350959300994873, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.049939028918743134, + "learning_rate": 0.000589856958339686, + "loss": 5.3317155838012695, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.05300058051943779, + "learning_rate": 0.0005897877680664465, + "loss": 5.353314399719238, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.05943184345960617, + "learning_rate": 0.0005897183467334822, + "loss": 5.3683929443359375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.05600789934396744, + "learning_rate": 0.0005896486943967249, + "loss": 5.345516204833984, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.05392718315124512, + "learning_rate": 0.0005895788111122919, + "loss": 5.291389465332031, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.037913475185632706, + "learning_rate": 0.0005895086969364871, + "loss": 5.353423118591309, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.03362123295664787, + "learning_rate": 0.0005894383519258001, + "loss": 5.379500389099121, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032558735460042953, + "learning_rate": 0.0005893677761369066, + "loss": 5.34043025970459, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.04074402526021004, + "learning_rate": 0.0005892969696266683, + "loss": 5.303869247436523, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.04270484298467636, + "learning_rate": 0.0005892259324521328, + "loss": 5.27297830581665, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.04604959115386009, + "learning_rate": 0.0005891546646705334, + "loss": 5.313268184661865, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.04462246224284172, + "learning_rate": 0.0005890831663392891, + "loss": 5.31167459487915, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.04966094344854355, + "learning_rate": 0.0005890114375160052, + "loss": 5.316695690155029, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.04662073776125908, + "learning_rate": 0.0005889394782584718, + "loss": 5.316898822784424, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.04380315542221069, + "learning_rate": 0.0005888672886246656, + "loss": 5.285087585449219, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03239579498767853, + "learning_rate": 0.0005887948686727483, + "loss": 5.304283142089844, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.029690183699131012, + "learning_rate": 0.0005887222184610675, + "loss": 5.305715084075928, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.03444967418909073, + "learning_rate": 0.0005886493380481559, + "loss": 5.2707672119140625, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.03302565962076187, + "learning_rate": 0.0005885762274927322, + "loss": 5.277251243591309, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.028477763757109642, + "learning_rate": 0.0005885028868537, + "loss": 5.295318603515625, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.022988801822066307, + "learning_rate": 0.0005884293161901487, + "loss": 5.249057292938232, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.026115216314792633, + "learning_rate": 0.000588355515561353, + "loss": 5.270636081695557, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.02576332725584507, + "learning_rate": 0.0005882814850267723, + "loss": 5.263096332550049, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.023978352546691895, + "learning_rate": 0.0005882072246460521, + "loss": 5.276605606079102, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.020695513114333153, + "learning_rate": 0.0005881327344790223, + "loss": 5.2141432762146, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01952219195663929, + "learning_rate": 0.0005880580145856986, + "loss": 5.214289665222168, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.022633206099271774, + "learning_rate": 0.0005879830650262813, + "loss": 5.236690998077393, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.023082347586750984, + "learning_rate": 0.0005879078858611557, + "loss": 5.2092156410217285, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.019620463252067566, + "learning_rate": 0.0005878324771508927, + "loss": 5.264451503753662, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.021504493430256844, + "learning_rate": 0.0005877568389562475, + "loss": 5.23444938659668, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.025672364979982376, + "learning_rate": 0.0005876809713381606, + "loss": 5.205289363861084, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.034436631947755814, + "learning_rate": 0.0005876048743577569, + "loss": 5.2047648429870605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.042389530688524246, + "learning_rate": 0.0005875285480763466, + "loss": 5.192008018493652, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.03310716897249222, + "learning_rate": 0.0005874519925554244, + "loss": 5.162734031677246, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.03029288910329342, + "learning_rate": 0.0005873752078566694, + "loss": 5.212151527404785, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04619310051202774, + "learning_rate": 0.0005872981940419459, + "loss": 5.215909004211426, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.05618087574839592, + "learning_rate": 0.0005872209511733027, + "loss": 5.210192680358887, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.03783707693219185, + "learning_rate": 0.0005871434793129726, + "loss": 5.201597690582275, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.03721104562282562, + "learning_rate": 0.0005870657785233735, + "loss": 5.188891410827637, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.033336155116558075, + "learning_rate": 0.0005869878488671075, + "loss": 5.1598615646362305, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.2592363357543945, + "eval_runtime": 57.3343, + "eval_samples_per_second": 42.592, + "eval_steps_per_second": 1.343, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03719741478562355, + "learning_rate": 0.0005869096904069611, + "loss": 5.1646623611450195, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.04888090118765831, + "learning_rate": 0.0005868313032059052, + "loss": 5.171182155609131, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.05073149502277374, + "learning_rate": 0.0005867526873270949, + "loss": 5.183949947357178, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.04187731072306633, + "learning_rate": 0.0005866738428338695, + "loss": 5.176133155822754, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.039445552974939346, + "learning_rate": 0.000586594769789753, + "loss": 5.148100852966309, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.035551708191633224, + "learning_rate": 0.0005865154682584524, + "loss": 5.151512145996094, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.031175289303064346, + "learning_rate": 0.0005864359383038602, + "loss": 5.102468967437744, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.03696789592504501, + "learning_rate": 0.000586356179990052, + "loss": 5.119902610778809, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.04836135357618332, + "learning_rate": 0.0005862761933812875, + "loss": 5.15183162689209, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.05206914246082306, + "learning_rate": 0.0005861959785420106, + "loss": 5.154870986938477, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.046368084847927094, + "learning_rate": 0.000586115535536849, + "loss": 5.130651473999023, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.07146432995796204, + "learning_rate": 0.000586034864430614, + "loss": 5.175804138183594, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.08809789270162582, + "learning_rate": 0.0005859539652883008, + "loss": 5.229716777801514, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.08116360008716583, + "learning_rate": 0.0005858728381750888, + "loss": 5.21910285949707, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.08282570540904999, + "learning_rate": 0.0005857914831563402, + "loss": 5.2336506843566895, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.07403547316789627, + "learning_rate": 0.0005857099002976016, + "loss": 5.251604080200195, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.07024004310369492, + "learning_rate": 0.0005856280896646026, + "loss": 5.267114162445068, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.056330062448978424, + "learning_rate": 0.0005855460513232567, + "loss": 5.201976299285889, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.051763955503702164, + "learning_rate": 0.0005854637853396606, + "loss": 5.1733198165893555, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.039900992065668106, + "learning_rate": 0.0005853812917800945, + "loss": 5.222941875457764, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.04187863692641258, + "learning_rate": 0.0005852985707110221, + "loss": 5.165033340454102, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.042645812034606934, + "learning_rate": 0.0005852156221990901, + "loss": 5.157554626464844, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.03856075555086136, + "learning_rate": 0.0005851324463111289, + "loss": 5.184629440307617, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.029387330636382103, + "learning_rate": 0.0005850490431141516, + "loss": 5.118739128112793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.027584781870245934, + "learning_rate": 0.0005849654126753545, + "loss": 5.100855350494385, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.023655924946069717, + "learning_rate": 0.0005848815550621175, + "loss": 5.154520034790039, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.024163050577044487, + "learning_rate": 0.0005847974703420028, + "loss": 5.105189323425293, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.020369920879602432, + "learning_rate": 0.0005847131585827564, + "loss": 5.113671779632568, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.021030737087130547, + "learning_rate": 0.0005846286198523061, + "loss": 5.0847649574279785, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.01729249767959118, + "learning_rate": 0.0005845438542187638, + "loss": 5.053225517272949, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.018704712390899658, + "learning_rate": 0.0005844588617504236, + "loss": 5.079106330871582, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.02049064077436924, + "learning_rate": 0.0005843736425157621, + "loss": 5.044185161590576, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.020736856386065483, + "learning_rate": 0.000584288196583439, + "loss": 5.041932106018066, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.016543276607990265, + "learning_rate": 0.0005842025240222966, + "loss": 5.032522201538086, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.018414106220006943, + "learning_rate": 0.0005841166249013598, + "loss": 5.037397384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.01611698977649212, + "learning_rate": 0.0005840304992898359, + "loss": 5.038179397583008, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.014310561120510101, + "learning_rate": 0.0005839441472571147, + "loss": 5.0146484375, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.015472214668989182, + "learning_rate": 0.0005838575688727685, + "loss": 5.060323715209961, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.016373401507735252, + "learning_rate": 0.0005837707642065518, + "loss": 5.032100677490234, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.018694665282964706, + "learning_rate": 0.000583683733328402, + "loss": 5.036993026733398, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.019977087154984474, + "learning_rate": 0.0005835964763084378, + "loss": 4.991037368774414, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.023741086944937706, + "learning_rate": 0.000583508993216961, + "loss": 5.043832778930664, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.025061562657356262, + "learning_rate": 0.0005834212841244548, + "loss": 5.016516208648682, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.020848384127020836, + "learning_rate": 0.0005833333491015851, + "loss": 4.986289978027344, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.015086354687809944, + "learning_rate": 0.0005832451882191995, + "loss": 4.9948625564575195, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.015096964314579964, + "learning_rate": 0.0005831568015483274, + "loss": 4.996342658996582, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016932200640439987, + "learning_rate": 0.0005830681891601807, + "loss": 4.998758316040039, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.019936930388212204, + "learning_rate": 0.0005829793511261525, + "loss": 4.959854602813721, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.027658682316541672, + "learning_rate": 0.0005828902875178183, + "loss": 4.968629837036133, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.03299275413155556, + "learning_rate": 0.0005828009984069347, + "loss": 4.994814872741699, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.033717650920152664, + "learning_rate": 0.0005827114838654406, + "loss": 4.976295471191406, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.02969924360513687, + "learning_rate": 0.0005826217439654562, + "loss": 4.969155788421631, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.02846827358007431, + "learning_rate": 0.0005825317787792831, + "loss": 4.989721298217773, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.03670600801706314, + "learning_rate": 0.0005824415883794049, + "loss": 4.988689422607422, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.045570675283670425, + "learning_rate": 0.0005823511728384863, + "loss": 4.974407196044922, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.04564111679792404, + "learning_rate": 0.0005822605322293733, + "loss": 4.966331481933594, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.04173210263252258, + "learning_rate": 0.0005821696666250937, + "loss": 4.925197601318359, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.04228242486715317, + "learning_rate": 0.0005820785760988559, + "loss": 4.965020656585693, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.05144782364368439, + "learning_rate": 0.0005819872607240503, + "loss": 5.0022149085998535, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.049400657415390015, + "learning_rate": 0.0005818957205742478, + "loss": 4.9635114669799805, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.043692708015441895, + "learning_rate": 0.0005818039557232005, + "loss": 4.977723121643066, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.04891742020845413, + "learning_rate": 0.0005817119662448421, + "loss": 4.996312141418457, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.035564981400966644, + "learning_rate": 0.0005816197522132866, + "loss": 4.940932273864746, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04679827764630318, + "learning_rate": 0.0005815273137028292, + "loss": 4.983131408691406, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.05489890277385712, + "learning_rate": 0.0005814346507879459, + "loss": 4.977827548980713, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04640809819102287, + "learning_rate": 0.0005813417635432937, + "loss": 4.954761505126953, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04439452290534973, + "learning_rate": 0.00058124865204371, + "loss": 4.914865970611572, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.04587096720933914, + "learning_rate": 0.0005811553163642131, + "loss": 4.949592590332031, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.044119734317064285, + "learning_rate": 0.0005810617565800019, + "loss": 4.956068992614746, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.043133001774549484, + "learning_rate": 0.0005809679727664559, + "loss": 4.993522644042969, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04315278306603432, + "learning_rate": 0.0005808739649991348, + "loss": 4.894291877746582, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04819472134113312, + "learning_rate": 0.0005807797333537792, + "loss": 4.958306789398193, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.023706436157227, + "eval_runtime": 57.4082, + "eval_samples_per_second": 42.537, + "eval_steps_per_second": 1.341, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.04403403028845787, + "learning_rate": 0.0005806852779063098, + "loss": 4.884971618652344, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.041993748396635056, + "learning_rate": 0.0005805905987328275, + "loss": 4.916353225708008, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.042302004992961884, + "learning_rate": 0.0005804956959096138, + "loss": 4.9012556076049805, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.044500596821308136, + "learning_rate": 0.0005804005695131299, + "loss": 4.915240287780762, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.03914736211299896, + "learning_rate": 0.0005803052196200177, + "loss": 4.885915756225586, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.0361703559756279, + "learning_rate": 0.0005802096463070988, + "loss": 4.829287528991699, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03432302549481392, + "learning_rate": 0.0005801138496513749, + "loss": 4.926834583282471, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.0317804180085659, + "learning_rate": 0.0005800178297300276, + "loss": 4.8707170486450195, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.035261351615190506, + "learning_rate": 0.0005799215866204183, + "loss": 4.897791862487793, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.03276067599654198, + "learning_rate": 0.0005798251204000886, + "loss": 4.864504814147949, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.034757066518068314, + "learning_rate": 0.0005797284311467594, + "loss": 4.890812873840332, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.033726081252098083, + "learning_rate": 0.0005796315189383316, + "loss": 4.831684112548828, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.03074726276099682, + "learning_rate": 0.0005795343838528855, + "loss": 4.8534111976623535, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02215893566608429, + "learning_rate": 0.0005794370259686811, + "loss": 4.8311238288879395, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.02132033370435238, + "learning_rate": 0.0005793394453641581, + "loss": 4.842617034912109, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020825816318392754, + "learning_rate": 0.0005792416421179351, + "loss": 4.84535026550293, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.019334357231855392, + "learning_rate": 0.0005791436163088106, + "loss": 4.839937686920166, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017372334375977516, + "learning_rate": 0.0005790453680157622, + "loss": 4.8115553855896, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.0190406683832407, + "learning_rate": 0.0005789468973179468, + "loss": 4.832542419433594, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.017372574657201767, + "learning_rate": 0.0005788482042947004, + "loss": 4.793632507324219, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.019171632826328278, + "learning_rate": 0.0005787492890255382, + "loss": 4.776987552642822, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.02109278179705143, + "learning_rate": 0.0005786501515901545, + "loss": 4.8216657638549805, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.026234494522213936, + "learning_rate": 0.0005785507920684225, + "loss": 4.8109049797058105, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.0357641875743866, + "learning_rate": 0.0005784512105403944, + "loss": 4.821109771728516, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.03627947345376015, + "learning_rate": 0.000578351407086301, + "loss": 4.810632228851318, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.032253146171569824, + "learning_rate": 0.0005782513817865527, + "loss": 4.7930521965026855, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.03402210772037506, + "learning_rate": 0.0005781511347217375, + "loss": 4.772500991821289, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.0321584977209568, + "learning_rate": 0.000578050665972623, + "loss": 4.781399726867676, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.030242707580327988, + "learning_rate": 0.0005779499756201549, + "loss": 4.802542686462402, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.02361336164176464, + "learning_rate": 0.0005778490637454576, + "loss": 4.771912574768066, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.028781302273273468, + "learning_rate": 0.000577747930429834, + "loss": 4.796924591064453, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.027042683213949203, + "learning_rate": 0.0005776465757547654, + "loss": 4.745218276977539, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.02715984545648098, + "learning_rate": 0.0005775449998019113, + "loss": 4.792944431304932, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.0314096100628376, + "learning_rate": 0.00057744320265311, + "loss": 4.7649993896484375, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.028749193996191025, + "learning_rate": 0.0005773411843903773, + "loss": 4.7980546951293945, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.02655309997498989, + "learning_rate": 0.0005772389450959075, + "loss": 4.746520519256592, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.027497397735714912, + "learning_rate": 0.000577136484852073, + "loss": 4.7490715980529785, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.0343393050134182, + "learning_rate": 0.000577033803741424, + "loss": 4.74436092376709, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.04444412514567375, + "learning_rate": 0.0005769309018466891, + "loss": 4.755973815917969, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.0437556728720665, + "learning_rate": 0.0005768277792507744, + "loss": 4.784368515014648, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0369921550154686, + "learning_rate": 0.0005767244360367638, + "loss": 4.753410339355469, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.05038141459226608, + "learning_rate": 0.0005766208722879192, + "loss": 4.755491256713867, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.04511422663927078, + "learning_rate": 0.00057651708808768, + "loss": 4.760293006896973, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.048762883991003036, + "learning_rate": 0.0005764130835196632, + "loss": 4.755958080291748, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.06749686598777771, + "learning_rate": 0.0005763088586676634, + "loss": 4.802557468414307, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.06451250612735748, + "learning_rate": 0.0005762044136156528, + "loss": 4.812925338745117, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.05436156690120697, + "learning_rate": 0.0005760997484477809, + "loss": 4.807103157043457, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.05233754590153694, + "learning_rate": 0.0005759948632483742, + "loss": 4.782449722290039, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.05519746616482735, + "learning_rate": 0.0005758897581019371, + "loss": 4.808449745178223, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.0578567236661911, + "learning_rate": 0.000575784433093151, + "loss": 4.7676005363464355, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.0588264986872673, + "learning_rate": 0.0005756788883068743, + "loss": 4.805377960205078, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04225790873169899, + "learning_rate": 0.0005755731238281423, + "loss": 4.748775005340576, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.0376368910074234, + "learning_rate": 0.0005754671397421678, + "loss": 4.759791374206543, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03830771893262863, + "learning_rate": 0.0005753609361343402, + "loss": 4.729974746704102, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.03684520348906517, + "learning_rate": 0.0005752545130902256, + "loss": 4.729854583740234, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02983408235013485, + "learning_rate": 0.0005751478706955674, + "loss": 4.78502082824707, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02854904904961586, + "learning_rate": 0.0005750410090362854, + "loss": 4.717472076416016, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.023819150403141975, + "learning_rate": 0.0005749339281984761, + "loss": 4.717124938964844, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.02477158047258854, + "learning_rate": 0.0005748266282684124, + "loss": 4.71781063079834, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.022002534940838814, + "learning_rate": 0.0005747191093325443, + "loss": 4.7480669021606445, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.02133483812212944, + "learning_rate": 0.0005746113714774976, + "loss": 4.693770408630371, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.022407138720154762, + "learning_rate": 0.0005745034147900747, + "loss": 4.720734119415283, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.020850926637649536, + "learning_rate": 0.0005743952393572544, + "loss": 4.745505332946777, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.020420582965016365, + "learning_rate": 0.0005742868452661918, + "loss": 4.688610076904297, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.019014107063412666, + "learning_rate": 0.0005741782326042181, + "loss": 4.673436164855957, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.017393039539456367, + "learning_rate": 0.0005740694014588403, + "loss": 4.6941118240356445, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.017877329140901566, + "learning_rate": 0.0005739603519177419, + "loss": 4.667474269866943, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.017749635502696037, + "learning_rate": 0.0005738510840687821, + "loss": 4.6901092529296875, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.018746376037597656, + "learning_rate": 0.000573741597999996, + "loss": 4.680166721343994, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016260217875242233, + "learning_rate": 0.0005736318937995947, + "loss": 4.6799821853637695, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015099814161658287, + "learning_rate": 0.0005735219715559646, + "loss": 4.680666923522949, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.016539735719561577, + "learning_rate": 0.0005734118313576683, + "loss": 4.639509201049805, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 4.768754482269287, + "eval_runtime": 58.3352, + "eval_samples_per_second": 41.862, + "eval_steps_per_second": 1.32, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.019733518362045288, + "learning_rate": 0.0005733014732934436, + "loss": 4.635312080383301, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.022571036592125893, + "learning_rate": 0.0005731908974522042, + "loss": 4.6041765213012695, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.022238552570343018, + "learning_rate": 0.0005730801039230389, + "loss": 4.648976802825928, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.02115180902183056, + "learning_rate": 0.0005729690927952123, + "loss": 4.623968124389648, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.021789928898215294, + "learning_rate": 0.0005728578641581637, + "loss": 4.620028018951416, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.02593887411057949, + "learning_rate": 0.0005727464181015081, + "loss": 4.591062545776367, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.0347842313349247, + "learning_rate": 0.0005726347547150357, + "loss": 4.632149696350098, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04119519516825676, + "learning_rate": 0.0005725228740887117, + "loss": 4.6254191398620605, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.04895750433206558, + "learning_rate": 0.0005724107763126761, + "loss": 4.621372222900391, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.05752161890268326, + "learning_rate": 0.0005722984614772442, + "loss": 4.61965274810791, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05115870013833046, + "learning_rate": 0.000572185929672906, + "loss": 4.63119649887085, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.05721089243888855, + "learning_rate": 0.0005720731809903263, + "loss": 4.652647018432617, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.0646839588880539, + "learning_rate": 0.000571960215520345, + "loss": 4.642184257507324, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.0682854950428009, + "learning_rate": 0.0005718470333539757, + "loss": 4.676589012145996, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.057402849197387695, + "learning_rate": 0.0005717336345824077, + "loss": 4.6585283279418945, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.05885373800992966, + "learning_rate": 0.0005716200192970043, + "loss": 4.671728134155273, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.05265705659985542, + "learning_rate": 0.0005715061875893032, + "loss": 4.641249656677246, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.05417842045426369, + "learning_rate": 0.0005713921395510166, + "loss": 4.661381721496582, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.06602369248867035, + "learning_rate": 0.0005712778752740307, + "loss": 4.684137344360352, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.06004182621836662, + "learning_rate": 0.0005711633948504066, + "loss": 4.625332832336426, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.04801511764526367, + "learning_rate": 0.0005710486983723787, + "loss": 4.664791107177734, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.04380105435848236, + "learning_rate": 0.0005709337859323561, + "loss": 4.664966583251953, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.03858727216720581, + "learning_rate": 0.0005708186576229218, + "loss": 4.661867618560791, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03240576758980751, + "learning_rate": 0.0005707033135368323, + "loss": 4.5855326652526855, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.028561925515532494, + "learning_rate": 0.0005705877537670184, + "loss": 4.637331008911133, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.026389246806502342, + "learning_rate": 0.0005704719784065846, + "loss": 4.598666667938232, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.02484819106757641, + "learning_rate": 0.0005703559875488088, + "loss": 4.601729393005371, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.02522977627813816, + "learning_rate": 0.0005702397812871429, + "loss": 4.609833717346191, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02589508146047592, + "learning_rate": 0.0005701233597152121, + "loss": 4.574592590332031, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.023374728858470917, + "learning_rate": 0.0005700067229268154, + "loss": 4.61362361907959, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.02465851791203022, + "learning_rate": 0.0005698898710159245, + "loss": 4.639899253845215, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.02258715033531189, + "learning_rate": 0.0005697728040766852, + "loss": 4.5656538009643555, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.018009789288043976, + "learning_rate": 0.0005696555222034162, + "loss": 4.61449670791626, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02121281623840332, + "learning_rate": 0.0005695380254906094, + "loss": 4.553830623626709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.018078701570630074, + "learning_rate": 0.0005694203140329296, + "loss": 4.534579277038574, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.0169303547590971, + "learning_rate": 0.000569302387925215, + "loss": 4.533995628356934, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.016318142414093018, + "learning_rate": 0.0005691842472624764, + "loss": 4.544438362121582, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.016664575785398483, + "learning_rate": 0.0005690658921398977, + "loss": 4.5913286209106445, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.01641504280269146, + "learning_rate": 0.0005689473226528354, + "loss": 4.520743370056152, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.017012735828757286, + "learning_rate": 0.0005688285388968187, + "loss": 4.517453193664551, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01672312617301941, + "learning_rate": 0.0005687095409675499, + "loss": 4.5235915184021, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.016893865540623665, + "learning_rate": 0.000568590328960903, + "loss": 4.543200492858887, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.017530135810375214, + "learning_rate": 0.0005684709029729253, + "loss": 4.54912805557251, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.018698599189519882, + "learning_rate": 0.0005683512630998361, + "loss": 4.522650718688965, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01984071172773838, + "learning_rate": 0.000568231409438027, + "loss": 4.546173095703125, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.023218289017677307, + "learning_rate": 0.0005681113420840619, + "loss": 4.5305094718933105, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.028968270868062973, + "learning_rate": 0.000567991061134677, + "loss": 4.521440505981445, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.032370273023843765, + "learning_rate": 0.0005678705666867805, + "loss": 4.54168701171875, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.033967748284339905, + "learning_rate": 0.0005677498588374524, + "loss": 4.538717269897461, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.028757929801940918, + "learning_rate": 0.0005676289376839452, + "loss": 4.517228603363037, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.029757825657725334, + "learning_rate": 0.0005675078033236827, + "loss": 4.510310173034668, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.03581003099679947, + "learning_rate": 0.0005673864558542605, + "loss": 4.506901264190674, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.03483985736966133, + "learning_rate": 0.0005672648953734462, + "loss": 4.549284934997559, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.031016027554869652, + "learning_rate": 0.000567143121979179, + "loss": 4.530084609985352, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.029373526573181152, + "learning_rate": 0.0005670211357695693, + "loss": 4.492218971252441, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.031661372631788254, + "learning_rate": 0.0005668989368428994, + "loss": 4.502824783325195, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.027251387014985085, + "learning_rate": 0.0005667765252976227, + "loss": 4.49193811416626, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02263176068663597, + "learning_rate": 0.0005666539012323639, + "loss": 4.512019157409668, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.02355228364467621, + "learning_rate": 0.0005665310647459189, + "loss": 4.508917331695557, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.023297371342778206, + "learning_rate": 0.0005664080159372551, + "loss": 4.497771263122559, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.019364114850759506, + "learning_rate": 0.0005662847549055107, + "loss": 4.51470947265625, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.023899901658296585, + "learning_rate": 0.0005661612817499947, + "loss": 4.48771858215332, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.02330983616411686, + "learning_rate": 0.0005660375965701872, + "loss": 4.478496551513672, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.027252234518527985, + "learning_rate": 0.0005659136994657392, + "loss": 4.497672080993652, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.030886363238096237, + "learning_rate": 0.0005657895905364723, + "loss": 4.49789571762085, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02515559457242489, + "learning_rate": 0.0005656652698823788, + "loss": 4.490771293640137, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020107125863432884, + "learning_rate": 0.0005655407376036217, + "loss": 4.481879234313965, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.022076716646552086, + "learning_rate": 0.0005654159938005345, + "loss": 4.494697570800781, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.025525815784931183, + "learning_rate": 0.0005652910385736207, + "loss": 4.517453193664551, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.03005894646048546, + "learning_rate": 0.0005651658720235548, + "loss": 4.453815460205078, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.030178502202033997, + "learning_rate": 0.0005650404942511812, + "loss": 4.493922233581543, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.029249394312500954, + "learning_rate": 0.0005649149053575141, + "loss": 4.453763008117676, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 4.607339382171631, + "eval_runtime": 57.9216, + "eval_samples_per_second": 42.16, + "eval_steps_per_second": 1.329, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.03362950682640076, + "learning_rate": 0.0005647891054437386, + "loss": 4.430700302124023, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.04223684221506119, + "learning_rate": 0.0005646630946112096, + "loss": 4.486915588378906, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.05192171037197113, + "learning_rate": 0.0005645368729614512, + "loss": 4.467541694641113, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.042525868862867355, + "learning_rate": 0.0005644104405961582, + "loss": 4.45252799987793, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.0406370609998703, + "learning_rate": 0.0005642837976171949, + "loss": 4.414548873901367, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.044476233422756195, + "learning_rate": 0.0005641569441265952, + "loss": 4.453492164611816, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.04687326401472092, + "learning_rate": 0.0005640298802265626, + "loss": 4.477725028991699, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.04011611267924309, + "learning_rate": 0.0005639026060194704, + "loss": 4.423356533050537, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.04971884936094284, + "learning_rate": 0.0005637751216078609, + "loss": 4.4853515625, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.05863234028220177, + "learning_rate": 0.0005636474270944461, + "loss": 4.466001033782959, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.057710036635398865, + "learning_rate": 0.0005635195225821072, + "loss": 4.466723442077637, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.05155074968934059, + "learning_rate": 0.0005633914081738945, + "loss": 4.4823198318481445, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.04535634070634842, + "learning_rate": 0.0005632630839730275, + "loss": 4.471219062805176, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.045147854834795, + "learning_rate": 0.0005631345500828946, + "loss": 4.476888179779053, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.03730619326233864, + "learning_rate": 0.0005630058066070533, + "loss": 4.428932189941406, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.034599728882312775, + "learning_rate": 0.0005628768536492299, + "loss": 4.47627592086792, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.032904356718063354, + "learning_rate": 0.0005627476913133193, + "loss": 4.422671794891357, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.03356337174773216, + "learning_rate": 0.0005626183197033856, + "loss": 4.456406593322754, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.035042595118284225, + "learning_rate": 0.0005624887389236609, + "loss": 4.371820449829102, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.03269032761454582, + "learning_rate": 0.0005623589490785462, + "loss": 4.402605056762695, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.028261031955480576, + "learning_rate": 0.0005622289502726108, + "loss": 4.433136940002441, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03301700949668884, + "learning_rate": 0.0005620987426105925, + "loss": 4.388359546661377, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.031664177775382996, + "learning_rate": 0.000561968326197397, + "loss": 4.41859245300293, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.026281699538230896, + "learning_rate": 0.0005618377011380988, + "loss": 4.406128883361816, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.023097911849617958, + "learning_rate": 0.0005617068675379398, + "loss": 4.389717102050781, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.022349286824464798, + "learning_rate": 0.0005615758255023306, + "loss": 4.408173561096191, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.021958226338028908, + "learning_rate": 0.0005614445751368491, + "loss": 4.434830188751221, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024551713839173317, + "learning_rate": 0.0005613131165472415, + "loss": 4.386944770812988, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.028048967942595482, + "learning_rate": 0.0005611814498394216, + "loss": 4.402279376983643, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02733907848596573, + "learning_rate": 0.0005610495751194708, + "loss": 4.403069496154785, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.020858559757471085, + "learning_rate": 0.0005609174924936384, + "loss": 4.406920909881592, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.021491149440407753, + "learning_rate": 0.0005607852020683409, + "loss": 4.385799407958984, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.021495329216122627, + "learning_rate": 0.0005606527039501621, + "loss": 4.4252495765686035, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.019708355888724327, + "learning_rate": 0.0005605199982458535, + "loss": 4.388864994049072, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.01886274851858616, + "learning_rate": 0.0005603870850623338, + "loss": 4.335962295532227, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.01955341547727585, + "learning_rate": 0.0005602539645066884, + "loss": 4.373560905456543, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.018962012603878975, + "learning_rate": 0.0005601206366861706, + "loss": 4.380022048950195, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.016663335263729095, + "learning_rate": 0.0005599871017081999, + "loss": 4.421034336090088, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01676873490214348, + "learning_rate": 0.0005598533596803631, + "loss": 4.382775783538818, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.014379126951098442, + "learning_rate": 0.0005597194107104137, + "loss": 4.41373872756958, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.015897583216428757, + "learning_rate": 0.000559585254906272, + "loss": 4.377284049987793, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.016513120383024216, + "learning_rate": 0.000559450892376025, + "loss": 4.41200065612793, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.016625918447971344, + "learning_rate": 0.000559316323227926, + "loss": 4.381187438964844, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.016028188169002533, + "learning_rate": 0.000559181547570395, + "loss": 4.353909015655518, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01682914048433304, + "learning_rate": 0.0005590465655120183, + "loss": 4.382326126098633, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.017078889533877373, + "learning_rate": 0.0005589113771615486, + "loss": 4.334949493408203, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.018182938918471336, + "learning_rate": 0.0005587759826279046, + "loss": 4.348130226135254, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01940886303782463, + "learning_rate": 0.0005586403820201713, + "loss": 4.339404106140137, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.020644692704081535, + "learning_rate": 0.0005585045754475995, + "loss": 4.358404159545898, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02433430217206478, + "learning_rate": 0.0005583685630196064, + "loss": 4.3660502433776855, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.027824537828564644, + "learning_rate": 0.0005582323448457744, + "loss": 4.34050178527832, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.02973582223057747, + "learning_rate": 0.0005580959210358523, + "loss": 4.348322868347168, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.030396128073334694, + "learning_rate": 0.0005579592916997542, + "loss": 4.356224060058594, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.029760530218482018, + "learning_rate": 0.0005578224569475599, + "loss": 4.357470989227295, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.02855013869702816, + "learning_rate": 0.0005576854168895147, + "loss": 4.335353374481201, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03376635164022446, + "learning_rate": 0.0005575481716360292, + "loss": 4.331794738769531, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.034431129693984985, + "learning_rate": 0.0005574107212976796, + "loss": 4.336880207061768, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.03524135425686836, + "learning_rate": 0.000557273065985207, + "loss": 4.31537389755249, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.03152058273553848, + "learning_rate": 0.0005571352058095179, + "loss": 4.37662410736084, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02878640592098236, + "learning_rate": 0.0005569971408816838, + "loss": 4.292388916015625, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.029663654044270515, + "learning_rate": 0.0005568588713129409, + "loss": 4.344346523284912, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.027895759791135788, + "learning_rate": 0.0005567203972146906, + "loss": 4.365072727203369, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.026465769857168198, + "learning_rate": 0.000556581718698499, + "loss": 4.334959506988525, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.03191852569580078, + "learning_rate": 0.000556442835876097, + "loss": 4.361069679260254, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.03535444661974907, + "learning_rate": 0.0005563037488593799, + "loss": 4.349765777587891, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03451581299304962, + "learning_rate": 0.0005561644577604074, + "loss": 4.343142509460449, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029810845851898193, + "learning_rate": 0.000556024962691404, + "loss": 4.307485580444336, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.027523217722773552, + "learning_rate": 0.0005558852637647584, + "loss": 4.366406440734863, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03043208085000515, + "learning_rate": 0.0005557453610930235, + "loss": 4.356929779052734, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.03058922104537487, + "learning_rate": 0.0005556052547889161, + "loss": 4.348942279815674, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.032052166759967804, + "learning_rate": 0.0005554649449653177, + "loss": 4.3496809005737305, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.03170694038271904, + "learning_rate": 0.0005553244317352731, + "loss": 4.315959930419922, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 4.4757890701293945, + "eval_runtime": 57.7451, + "eval_samples_per_second": 42.289, + "eval_steps_per_second": 1.333, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.029025450348854065, + "learning_rate": 0.0005551837152119915, + "loss": 4.288647174835205, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.029403962194919586, + "learning_rate": 0.0005550427955088455, + "loss": 4.288555145263672, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.03265019878745079, + "learning_rate": 0.0005549016727393715, + "loss": 4.292625904083252, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.0387202650308609, + "learning_rate": 0.0005547603470172697, + "loss": 4.238926887512207, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.04554206505417824, + "learning_rate": 0.0005546188184564036, + "loss": 4.29966402053833, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.036231476813554764, + "learning_rate": 0.0005544770871708004, + "loss": 4.304398536682129, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.038703009486198425, + "learning_rate": 0.0005543351532746501, + "loss": 4.2945356369018555, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03224619850516319, + "learning_rate": 0.0005541930168823067, + "loss": 4.261351585388184, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.029722856357693672, + "learning_rate": 0.0005540506781082864, + "loss": 4.303143501281738, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.02764045260846615, + "learning_rate": 0.0005539081370672695, + "loss": 4.243679523468018, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.0312200877815485, + "learning_rate": 0.0005537653938740984, + "loss": 4.279904365539551, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.027349065989255905, + "learning_rate": 0.0005536224486437788, + "loss": 4.270228385925293, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.030385233461856842, + "learning_rate": 0.0005534793014914793, + "loss": 4.301590442657471, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.03292274847626686, + "learning_rate": 0.0005533359525325307, + "loss": 4.301828384399414, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.03265977278351784, + "learning_rate": 0.0005531924018824266, + "loss": 4.286200523376465, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.03177836537361145, + "learning_rate": 0.0005530486496568236, + "loss": 4.296319007873535, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.028879934921860695, + "learning_rate": 0.0005529046959715397, + "loss": 4.261706829071045, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.02855812944471836, + "learning_rate": 0.0005527605409425562, + "loss": 4.273903846740723, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03037901036441326, + "learning_rate": 0.0005526161846860161, + "loss": 4.248080253601074, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.029279105365276337, + "learning_rate": 0.0005524716273182245, + "loss": 4.281311988830566, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.02592580020427704, + "learning_rate": 0.0005523268689556489, + "loss": 4.204848289489746, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.021899603307247162, + "learning_rate": 0.0005521819097149183, + "loss": 4.232453346252441, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.019704442471265793, + "learning_rate": 0.0005520367497128238, + "loss": 4.253462314605713, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.019550519064068794, + "learning_rate": 0.0005518913890663182, + "loss": 4.307737350463867, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.020320113748311996, + "learning_rate": 0.0005517458278925161, + "loss": 4.257411956787109, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02118559181690216, + "learning_rate": 0.0005516000663086932, + "loss": 4.262089729309082, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.02050616219639778, + "learning_rate": 0.0005514541044322872, + "loss": 4.2490644454956055, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019793255254626274, + "learning_rate": 0.0005513079423808967, + "loss": 4.267400741577148, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.020760254934430122, + "learning_rate": 0.0005511615802722823, + "loss": 4.2532429695129395, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.019524799659848213, + "learning_rate": 0.0005510150182243647, + "loss": 4.267422199249268, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.020166637375950813, + "learning_rate": 0.0005508682563552269, + "loss": 4.214296817779541, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.01964394748210907, + "learning_rate": 0.0005507212947831119, + "loss": 4.252087116241455, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01690433733165264, + "learning_rate": 0.0005505741336264239, + "loss": 4.232336044311523, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.017682762816548347, + "learning_rate": 0.0005504267730037283, + "loss": 4.213962554931641, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01688133366405964, + "learning_rate": 0.0005502792130337508, + "loss": 4.199461460113525, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.019166061654686928, + "learning_rate": 0.0005501314538353776, + "loss": 4.228936672210693, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.024065518751740456, + "learning_rate": 0.000549983495527656, + "loss": 4.214227676391602, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02518288791179657, + "learning_rate": 0.000549835338229793, + "loss": 4.256412029266357, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.026093967258930206, + "learning_rate": 0.0005496869820611563, + "loss": 4.216564178466797, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.023775635287165642, + "learning_rate": 0.0005495384271412739, + "loss": 4.226430892944336, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.018810907378792763, + "learning_rate": 0.0005493896735898338, + "loss": 4.190013885498047, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.018974872305989265, + "learning_rate": 0.000549240721526684, + "loss": 4.246231555938721, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.02069833129644394, + "learning_rate": 0.0005490915710718325, + "loss": 4.198457717895508, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.02097306028008461, + "learning_rate": 0.0005489422223454471, + "loss": 4.212507247924805, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022165542468428612, + "learning_rate": 0.0005487926754678556, + "loss": 4.2043843269348145, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.019068919122219086, + "learning_rate": 0.0005486429305595448, + "loss": 4.197460651397705, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01701812446117401, + "learning_rate": 0.0005484929877411618, + "loss": 4.220056533813477, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01629726029932499, + "learning_rate": 0.0005483428471335127, + "loss": 4.214868545532227, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018661806359887123, + "learning_rate": 0.000548192508857563, + "loss": 4.227047920227051, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.02005891129374504, + "learning_rate": 0.0005480419730344377, + "loss": 4.197265625, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.02591128647327423, + "learning_rate": 0.0005478912397854208, + "loss": 4.211041450500488, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.032659754157066345, + "learning_rate": 0.0005477403092319552, + "loss": 4.230195045471191, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.04057808965444565, + "learning_rate": 0.0005475891814956429, + "loss": 4.223843097686768, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.04103454574942589, + "learning_rate": 0.0005474378566982449, + "loss": 4.197145462036133, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.03810729831457138, + "learning_rate": 0.0005472863349616809, + "loss": 4.237957000732422, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.028838301077485085, + "learning_rate": 0.0005471346164080292, + "loss": 4.206722259521484, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02436191774904728, + "learning_rate": 0.0005469827011595266, + "loss": 4.1983137130737305, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.022839387878775597, + "learning_rate": 0.0005468305893385686, + "loss": 4.20530366897583, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.02375260926783085, + "learning_rate": 0.0005466782810677088, + "loss": 4.213202476501465, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.02177741937339306, + "learning_rate": 0.0005465257764696593, + "loss": 4.200539588928223, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.02351468615233898, + "learning_rate": 0.0005463730756672903, + "loss": 4.198091506958008, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.023375794291496277, + "learning_rate": 0.0005462201787836301, + "loss": 4.193244934082031, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.02479805052280426, + "learning_rate": 0.0005460670859418651, + "loss": 4.2250871658325195, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.026894841343164444, + "learning_rate": 0.000545913797265339, + "loss": 4.241791248321533, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.030582403764128685, + "learning_rate": 0.0005457603128775541, + "loss": 4.203666687011719, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.032222259789705276, + "learning_rate": 0.0005456066329021699, + "loss": 4.209629535675049, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.03181060031056404, + "learning_rate": 0.0005454527574630037, + "loss": 4.200459957122803, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02687576413154602, + "learning_rate": 0.0005452986866840298, + "loss": 4.211819648742676, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.025988668203353882, + "learning_rate": 0.0005451444206893806, + "loss": 4.198040962219238, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.02937682718038559, + "learning_rate": 0.0005449899596033453, + "loss": 4.222935676574707, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.03043951839208603, + "learning_rate": 0.0005448353035503704, + "loss": 4.2031755447387695, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.030401621013879776, + "learning_rate": 0.0005446804526550596, + "loss": 4.198380470275879, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 4.368444919586182, + "eval_runtime": 58.3009, + "eval_samples_per_second": 41.886, + "eval_steps_per_second": 1.321, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.03461585193872452, + "learning_rate": 0.0005445254070421732, + "loss": 4.147369384765625, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.039084456861019135, + "learning_rate": 0.0005443701668366288, + "loss": 4.124129295349121, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.04021848738193512, + "learning_rate": 0.0005442147321635007, + "loss": 4.127362251281738, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.041000768542289734, + "learning_rate": 0.00054405910314802, + "loss": 4.1432061195373535, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.04008613899350166, + "learning_rate": 0.0005439032799155737, + "loss": 4.169931888580322, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.039394307881593704, + "learning_rate": 0.000543747262591706, + "loss": 4.155813217163086, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.04374868795275688, + "learning_rate": 0.0005435910513021172, + "loss": 4.159891128540039, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0443582646548748, + "learning_rate": 0.0005434346461726641, + "loss": 4.157362937927246, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.04601646214723587, + "learning_rate": 0.0005432780473293593, + "loss": 4.152097225189209, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.04327242821455002, + "learning_rate": 0.0005431212548983718, + "loss": 4.170074462890625, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.0413462370634079, + "learning_rate": 0.0005429642690060262, + "loss": 4.156710147857666, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.036855220794677734, + "learning_rate": 0.0005428070897788034, + "loss": 4.138053894042969, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.03781217709183693, + "learning_rate": 0.0005426497173433398, + "loss": 4.172306060791016, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.03886301815509796, + "learning_rate": 0.0005424921518264275, + "loss": 4.135544776916504, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.03730994835495949, + "learning_rate": 0.0005423343933550143, + "loss": 4.171654224395752, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.03271950036287308, + "learning_rate": 0.0005421764420562032, + "loss": 4.169906139373779, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.03292673081159592, + "learning_rate": 0.0005420182980572527, + "loss": 4.163660049438477, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.029366906732320786, + "learning_rate": 0.0005418599614855768, + "loss": 4.161055088043213, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.023805437609553337, + "learning_rate": 0.0005417014324687442, + "loss": 4.172390937805176, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.023178137838840485, + "learning_rate": 0.000541542711134479, + "loss": 4.146788597106934, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.02403450198471546, + "learning_rate": 0.0005413837976106599, + "loss": 4.124364852905273, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.025009717792272568, + "learning_rate": 0.0005412246920253208, + "loss": 4.12275505065918, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.023431289941072464, + "learning_rate": 0.0005410653945066503, + "loss": 4.134739398956299, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.023803815245628357, + "learning_rate": 0.0005409059051829914, + "loss": 4.130463123321533, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.021219994872808456, + "learning_rate": 0.0005407462241828417, + "loss": 4.124577522277832, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.01899666152894497, + "learning_rate": 0.0005405863516348535, + "loss": 4.167086124420166, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.01889493502676487, + "learning_rate": 0.000540426287667833, + "loss": 4.152744293212891, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.019546914845705032, + "learning_rate": 0.000540266032410741, + "loss": 4.122655868530273, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02085448056459427, + "learning_rate": 0.0005401055859926923, + "loss": 4.141949653625488, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.019959403201937675, + "learning_rate": 0.0005399449485429555, + "loss": 4.131907939910889, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.01885392889380455, + "learning_rate": 0.0005397841201909535, + "loss": 4.118953227996826, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.018216131255030632, + "learning_rate": 0.0005396231010662627, + "loss": 4.096459865570068, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.019385207444429398, + "learning_rate": 0.0005394618912986132, + "loss": 4.111050128936768, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.018676603212952614, + "learning_rate": 0.0005393004910178889, + "loss": 4.156215667724609, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.01876339502632618, + "learning_rate": 0.0005391389003541271, + "loss": 4.1107282638549805, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.01831623539328575, + "learning_rate": 0.0005389771194375185, + "loss": 4.154116630554199, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.018896661698818207, + "learning_rate": 0.0005388151483984071, + "loss": 4.093320846557617, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.019582601264119148, + "learning_rate": 0.0005386529873672898, + "loss": 4.106500625610352, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.020313601940870285, + "learning_rate": 0.0005384906364748172, + "loss": 4.071961879730225, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.020385954529047012, + "learning_rate": 0.000538328095851792, + "loss": 4.129006862640381, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02231123298406601, + "learning_rate": 0.0005381653656291707, + "loss": 4.067756652832031, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.022957179695367813, + "learning_rate": 0.0005380024459380619, + "loss": 4.088230133056641, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021438462659716606, + "learning_rate": 0.000537839336909727, + "loss": 4.128036022186279, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01893841102719307, + "learning_rate": 0.00053767603867558, + "loss": 4.155407905578613, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.018539676442742348, + "learning_rate": 0.0005375125513671874, + "loss": 4.101880073547363, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.01881185919046402, + "learning_rate": 0.0005373488751162681, + "loss": 4.118906021118164, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.021657396107912064, + "learning_rate": 0.0005371850100546929, + "loss": 4.091935157775879, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.023293187841773033, + "learning_rate": 0.000537020956314485, + "loss": 4.105803489685059, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.021306317299604416, + "learning_rate": 0.0005368567140278196, + "loss": 4.0610809326171875, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.022078553214669228, + "learning_rate": 0.0005366922833270235, + "loss": 4.09867525100708, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.02266664244234562, + "learning_rate": 0.0005365276643445757, + "loss": 4.1027326583862305, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.02091757394373417, + "learning_rate": 0.0005363628572131069, + "loss": 4.080266952514648, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.020374659448862076, + "learning_rate": 0.0005361978620653991, + "loss": 4.088479042053223, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.02294750325381756, + "learning_rate": 0.0005360326790343858, + "loss": 4.084364891052246, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.02337353676557541, + "learning_rate": 0.000535867308253152, + "loss": 4.125696182250977, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.022643696516752243, + "learning_rate": 0.0005357017498549341, + "loss": 4.097160339355469, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.021080605685710907, + "learning_rate": 0.0005355360039731196, + "loss": 4.102363109588623, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.02269427664577961, + "learning_rate": 0.0005353700707412466, + "loss": 4.089844703674316, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.02195413038134575, + "learning_rate": 0.0005352039502930048, + "loss": 4.132293701171875, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.01805616170167923, + "learning_rate": 0.0005350376427622343, + "loss": 4.0842814445495605, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.018048783764243126, + "learning_rate": 0.0005348711482829261, + "loss": 4.109591007232666, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.0199508648365736, + "learning_rate": 0.000534704466989222, + "loss": 4.0908074378967285, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.020186785608530045, + "learning_rate": 0.0005345375990154138, + "loss": 4.077959060668945, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.022526193410158157, + "learning_rate": 0.0005343705444959441, + "loss": 4.109979629516602, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.025224441662430763, + "learning_rate": 0.000534203303565406, + "loss": 4.047533988952637, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.027292724698781967, + "learning_rate": 0.000534035876358542, + "loss": 4.09566593170166, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.025669090449810028, + "learning_rate": 0.0005338682630102454, + "loss": 4.0606160163879395, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02791150100529194, + "learning_rate": 0.0005337004636555595, + "loss": 4.097081184387207, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.03294133022427559, + "learning_rate": 0.0005335324784296768, + "loss": 4.121442794799805, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.027322066947817802, + "learning_rate": 0.0005333643074679403, + "loss": 4.1188740730285645, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.024866372346878052, + "learning_rate": 0.000533195950905842, + "loss": 4.082042694091797, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.02901027724146843, + "learning_rate": 0.0005330274088790239, + "loss": 4.110199451446533, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 4.283970355987549, + "eval_runtime": 58.4003, + "eval_samples_per_second": 41.815, + "eval_steps_per_second": 1.318, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.028958842158317566, + "learning_rate": 0.0005328586815232773, + "loss": 4.040733337402344, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.03491180017590523, + "learning_rate": 0.0005326897689745428, + "loss": 4.0473432540893555, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.04025818780064583, + "learning_rate": 0.0005325206713689101, + "loss": 4.021976470947266, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03998899459838867, + "learning_rate": 0.0005323513888426181, + "loss": 4.04901123046875, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.042165111750364304, + "learning_rate": 0.0005321819215320546, + "loss": 4.050230979919434, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.03511597216129303, + "learning_rate": 0.0005320122695737565, + "loss": 4.0989990234375, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.04064979776740074, + "learning_rate": 0.0005318424331044094, + "loss": 4.051448822021484, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.047536592930555344, + "learning_rate": 0.0005316724122608473, + "loss": 4.064278602600098, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.05592669919133186, + "learning_rate": 0.0005315022071800527, + "loss": 4.045516014099121, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.04262633994221687, + "learning_rate": 0.0005313318179991574, + "loss": 4.060666084289551, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03997781500220299, + "learning_rate": 0.0005311612448554402, + "loss": 4.065315246582031, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.038152433931827545, + "learning_rate": 0.0005309904878863293, + "loss": 4.069721221923828, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.0367623008787632, + "learning_rate": 0.0005308195472294001, + "loss": 4.080986976623535, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.0358598455786705, + "learning_rate": 0.0005306484230223769, + "loss": 4.065290927886963, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.039188165217638016, + "learning_rate": 0.000530477115403131, + "loss": 4.024333477020264, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.036839909851551056, + "learning_rate": 0.0005303056245096817, + "loss": 4.067532539367676, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.03456985577940941, + "learning_rate": 0.0005301339504801967, + "loss": 4.048443794250488, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.039487943053245544, + "learning_rate": 0.0005299620934529902, + "loss": 4.0585737228393555, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.03415827825665474, + "learning_rate": 0.0005297900535665244, + "loss": 4.06104850769043, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.03209620714187622, + "learning_rate": 0.0005296178309594087, + "loss": 4.009165287017822, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.028947360813617706, + "learning_rate": 0.0005294454257704, + "loss": 4.067384243011475, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026836825534701347, + "learning_rate": 0.0005292728381384018, + "loss": 4.046419620513916, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024095216765999794, + "learning_rate": 0.0005291000682024647, + "loss": 4.019008636474609, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.020965851843357086, + "learning_rate": 0.0005289271161017867, + "loss": 4.020613670349121, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.02250855043530464, + "learning_rate": 0.0005287539819757119, + "loss": 4.031634330749512, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.021385207772254944, + "learning_rate": 0.0005285806659637315, + "loss": 4.086246013641357, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.021104589104652405, + "learning_rate": 0.000528407168205483, + "loss": 4.033970355987549, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.019584037363529205, + "learning_rate": 0.0005282334888407503, + "loss": 4.025118827819824, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.02107258513569832, + "learning_rate": 0.000528059628009464, + "loss": 4.020823955535889, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.020438389852643013, + "learning_rate": 0.0005278855858517003, + "loss": 4.016323566436768, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.01881488971412182, + "learning_rate": 0.0005277113625076822, + "loss": 3.9991841316223145, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.01763618178665638, + "learning_rate": 0.000527536958117778, + "loss": 4.0062150955200195, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.017308862879872322, + "learning_rate": 0.0005273623728225022, + "loss": 3.946723461151123, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01706845872104168, + "learning_rate": 0.0005271876067625153, + "loss": 4.056562900543213, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.018664294853806496, + "learning_rate": 0.0005270126600786228, + "loss": 4.024652481079102, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.018523404374718666, + "learning_rate": 0.0005268375329117762, + "loss": 3.962064743041992, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.017408935353159904, + "learning_rate": 0.0005266622254030724, + "loss": 4.008697509765625, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.01588570512831211, + "learning_rate": 0.0005264867376937534, + "loss": 4.011664390563965, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.015084310434758663, + "learning_rate": 0.0005263110699252065, + "loss": 3.996224880218506, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01637054979801178, + "learning_rate": 0.000526135222238964, + "loss": 4.015972137451172, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.01673802360892296, + "learning_rate": 0.0005259591947767032, + "loss": 3.975414276123047, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.016544019803404808, + "learning_rate": 0.0005257829876802464, + "loss": 4.012955188751221, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.017455460503697395, + "learning_rate": 0.0005256066010915603, + "loss": 4.010180473327637, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.017625460401177406, + "learning_rate": 0.0005254300351527566, + "loss": 4.001679420471191, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.01732795126736164, + "learning_rate": 0.000525253290006091, + "loss": 4.006107330322266, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.016654646024107933, + "learning_rate": 0.0005250763657939641, + "loss": 4.045180797576904, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.015085658989846706, + "learning_rate": 0.0005248992626589203, + "loss": 3.994246482849121, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.015112568624317646, + "learning_rate": 0.0005247219807436489, + "loss": 4.004963397979736, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.014882124960422516, + "learning_rate": 0.000524544520190982, + "loss": 4.029565811157227, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.016012471169233322, + "learning_rate": 0.000524366881143897, + "loss": 4.018208026885986, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.015761924907565117, + "learning_rate": 0.0005241890637455141, + "loss": 3.9911205768585205, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.017770355567336082, + "learning_rate": 0.0005240110681390978, + "loss": 4.028733253479004, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01959751546382904, + "learning_rate": 0.0005238328944680558, + "loss": 4.003322124481201, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.018713824450969696, + "learning_rate": 0.0005236545428759394, + "loss": 3.991718292236328, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.01992208883166313, + "learning_rate": 0.0005234760135064434, + "loss": 4.002507209777832, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.021671291440725327, + "learning_rate": 0.0005232973065034055, + "loss": 4.021344184875488, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024282798171043396, + "learning_rate": 0.0005231184220108069, + "loss": 3.9899158477783203, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.023294344544410706, + "learning_rate": 0.0005229393601727713, + "loss": 4.046627044677734, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.02029707282781601, + "learning_rate": 0.000522760121133566, + "loss": 4.022454261779785, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.02151239849627018, + "learning_rate": 0.0005225807050376004, + "loss": 4.005476951599121, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.02320881187915802, + "learning_rate": 0.0005224011120294269, + "loss": 4.012726783752441, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.026484191417694092, + "learning_rate": 0.0005222213422537401, + "loss": 3.9904391765594482, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.026279661804437637, + "learning_rate": 0.0005220413958553777, + "loss": 3.9738848209381104, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.024299729615449905, + "learning_rate": 0.0005218612729793188, + "loss": 4.003316402435303, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.02337595447897911, + "learning_rate": 0.0005216809737706856, + "loss": 3.954984188079834, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.02416428178548813, + "learning_rate": 0.0005215004983747415, + "loss": 4.0317063331604, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.022488107904791832, + "learning_rate": 0.0005213198469368925, + "loss": 3.9736647605895996, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.019700225442647934, + "learning_rate": 0.0005211390196026861, + "loss": 3.979867458343506, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.020382333546876907, + "learning_rate": 0.0005209580165178117, + "loss": 3.995251178741455, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.020149298012256622, + "learning_rate": 0.0005207768378281003, + "loss": 3.9987268447875977, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.019790509715676308, + "learning_rate": 0.0005205954836795241, + "loss": 3.9838991165161133, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.022096451371908188, + "learning_rate": 0.000520413954218197, + "loss": 3.995800256729126, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.214347839355469, + "eval_runtime": 58.8872, + "eval_samples_per_second": 41.469, + "eval_steps_per_second": 1.308, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02203402668237686, + "learning_rate": 0.0005202322495903739, + "loss": 3.915332078933716, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.02336675301194191, + "learning_rate": 0.0005200503699424509, + "loss": 3.914801597595215, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.027339637279510498, + "learning_rate": 0.0005198683154209654, + "loss": 3.9277780055999756, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.026872653514146805, + "learning_rate": 0.0005196860861725952, + "loss": 3.9160094261169434, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.026558255776762962, + "learning_rate": 0.0005195036823441592, + "loss": 3.9084300994873047, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.02644367143511772, + "learning_rate": 0.0005193211040826169, + "loss": 3.9328718185424805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.024677276611328125, + "learning_rate": 0.0005191383515350681, + "loss": 3.88993501663208, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.023194244131445885, + "learning_rate": 0.0005189554248487536, + "loss": 3.923567056655884, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.024653147906064987, + "learning_rate": 0.0005187723241710539, + "loss": 3.8740999698638916, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.026041202247142792, + "learning_rate": 0.0005185890496494899, + "loss": 3.9542148113250732, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.029586758464574814, + "learning_rate": 0.0005184056014317225, + "loss": 3.9241347312927246, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.0325787179172039, + "learning_rate": 0.0005182219796655528, + "loss": 3.9656593799591064, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.03183380514383316, + "learning_rate": 0.0005180381844989214, + "loss": 3.950648307800293, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.03434471786022186, + "learning_rate": 0.0005178542160799088, + "loss": 3.9590439796447754, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.030099432915449142, + "learning_rate": 0.0005176700745567353, + "loss": 3.9518113136291504, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.025950564071536064, + "learning_rate": 0.0005174857600777599, + "loss": 3.8908939361572266, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.029959935694932938, + "learning_rate": 0.0005173012727914816, + "loss": 3.945037364959717, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.029551763087511063, + "learning_rate": 0.0005171166128465389, + "loss": 3.923959732055664, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.03538523241877556, + "learning_rate": 0.0005169317803917083, + "loss": 3.9395527839660645, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.037924155592918396, + "learning_rate": 0.0005167467755759065, + "loss": 3.945115089416504, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.036773014813661575, + "learning_rate": 0.0005165615985481881, + "loss": 3.9484646320343018, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.04125838354229927, + "learning_rate": 0.0005163762494577474, + "loss": 4.000699043273926, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.044125135987997055, + "learning_rate": 0.0005161907284539165, + "loss": 3.9828028678894043, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.040853384882211685, + "learning_rate": 0.0005160050356861663, + "loss": 3.957505702972412, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.03501008450984955, + "learning_rate": 0.0005158191713041063, + "loss": 3.9360849857330322, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.04082556813955307, + "learning_rate": 0.0005156331354574838, + "loss": 3.95941424369812, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.045432962477207184, + "learning_rate": 0.0005154469282961847, + "loss": 3.9666922092437744, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.04048244655132294, + "learning_rate": 0.0005152605499702328, + "loss": 3.984555244445801, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.031016314402222633, + "learning_rate": 0.0005150740006297896, + "loss": 3.945204257965088, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.03210771828889847, + "learning_rate": 0.0005148872804251548, + "loss": 3.976186752319336, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.03130023553967476, + "learning_rate": 0.000514700389506765, + "loss": 3.915727138519287, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.031193193048238754, + "learning_rate": 0.0005145133280251954, + "loss": 3.961866855621338, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.027518954128026962, + "learning_rate": 0.0005143260961311578, + "loss": 3.9742956161499023, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.02751155197620392, + "learning_rate": 0.0005141386939755014, + "loss": 3.949662208557129, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.026033353060483932, + "learning_rate": 0.0005139511217092129, + "loss": 3.9327926635742188, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.02579621411859989, + "learning_rate": 0.000513763379483416, + "loss": 3.9334418773651123, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.023009559139609337, + "learning_rate": 0.0005135754674493709, + "loss": 3.969707489013672, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.02196447364985943, + "learning_rate": 0.0005133873857584753, + "loss": 3.9084434509277344, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.02206484042108059, + "learning_rate": 0.0005131991345622629, + "loss": 3.922117233276367, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.023360205814242363, + "learning_rate": 0.0005130107140124046, + "loss": 3.965158224105835, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02162669226527214, + "learning_rate": 0.0005128221242607072, + "loss": 3.9519729614257812, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.019719846546649933, + "learning_rate": 0.0005126333654591141, + "loss": 3.9195961952209473, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.018988260999321938, + "learning_rate": 0.0005124444377597049, + "loss": 3.937926769256592, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.019278401508927345, + "learning_rate": 0.0005122553413146951, + "loss": 3.9126808643341064, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019101005047559738, + "learning_rate": 0.0005120660762764366, + "loss": 3.8993749618530273, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.019156700000166893, + "learning_rate": 0.0005118766427974166, + "loss": 3.9531068801879883, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019944407045841217, + "learning_rate": 0.0005116870410302582, + "loss": 3.9645180702209473, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.01826174184679985, + "learning_rate": 0.0005114972711277202, + "loss": 3.948477268218994, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.0191368218511343, + "learning_rate": 0.0005113073332426968, + "loss": 3.944648027420044, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.017768343910574913, + "learning_rate": 0.0005111172275282174, + "loss": 3.9584474563598633, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.01955590397119522, + "learning_rate": 0.0005109269541374469, + "loss": 3.8936800956726074, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.02100536972284317, + "learning_rate": 0.000510736513223685, + "loss": 3.93919038772583, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.020873935893177986, + "learning_rate": 0.0005105459049403665, + "loss": 3.929244041442871, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.021876046434044838, + "learning_rate": 0.0005103551294410614, + "loss": 3.9264731407165527, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.021726684644818306, + "learning_rate": 0.0005101641868794736, + "loss": 3.923135280609131, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.019308188930153847, + "learning_rate": 0.0005099730774094423, + "loss": 3.9062135219573975, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01917017623782158, + "learning_rate": 0.0005097818011849408, + "loss": 3.898810386657715, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017933931201696396, + "learning_rate": 0.0005095903583600772, + "loss": 3.8899734020233154, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.015131795778870583, + "learning_rate": 0.0005093987490890933, + "loss": 3.9452900886535645, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.017190825194120407, + "learning_rate": 0.0005092069735263651, + "loss": 3.9124746322631836, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.01618487387895584, + "learning_rate": 0.000509015031826403, + "loss": 3.91654109954834, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.014920198358595371, + "learning_rate": 0.0005088229241438509, + "loss": 3.9144535064697266, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01615205965936184, + "learning_rate": 0.0005086306506334864, + "loss": 3.917003631591797, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.015370192006230354, + "learning_rate": 0.0005084382114502208, + "loss": 3.8910603523254395, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.015290996059775352, + "learning_rate": 0.0005082456067490988, + "loss": 3.9122695922851562, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.015913106501102448, + "learning_rate": 0.0005080528366852987, + "loss": 3.9304556846618652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.01639600843191147, + "learning_rate": 0.0005078599014141318, + "loss": 3.8974502086639404, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.017010601237416267, + "learning_rate": 0.0005076668010910425, + "loss": 3.9723148345947266, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.017778996378183365, + "learning_rate": 0.0005074735358716083, + "loss": 3.8905391693115234, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.020029611885547638, + "learning_rate": 0.0005072801059115394, + "loss": 3.934976100921631, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.01931857503950596, + "learning_rate": 0.000507086511366679, + "loss": 3.9152774810791016, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019750485196709633, + "learning_rate": 0.0005068927523930027, + "loss": 3.877650499343872, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.161914825439453, + "eval_runtime": 58.5217, + "eval_samples_per_second": 41.728, + "eval_steps_per_second": 1.316, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02073049545288086, + "learning_rate": 0.0005066988291466184, + "loss": 3.8331246376037598, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02272389829158783, + "learning_rate": 0.0005065047417837669, + "loss": 3.8847155570983887, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02139078453183174, + "learning_rate": 0.0005063104904608206, + "loss": 3.8276169300079346, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02043229155242443, + "learning_rate": 0.0005061160753342843, + "loss": 3.8804001808166504, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.02345564216375351, + "learning_rate": 0.0005059214965607948, + "loss": 3.851609468460083, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.023448185995221138, + "learning_rate": 0.0005057267542971207, + "loss": 3.8446507453918457, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.024826066568493843, + "learning_rate": 0.0005055318487001624, + "loss": 3.86712384223938, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.02491980977356434, + "learning_rate": 0.0005053367799269518, + "loss": 3.8374528884887695, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.023002834990620613, + "learning_rate": 0.0005051415481346522, + "loss": 3.783214569091797, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.02123952843248844, + "learning_rate": 0.0005049461534805585, + "loss": 3.8463428020477295, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.01981225423514843, + "learning_rate": 0.0005047505961220964, + "loss": 3.8307082653045654, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01915367692708969, + "learning_rate": 0.0005045548762168231, + "loss": 3.8190226554870605, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.01890859752893448, + "learning_rate": 0.0005043589939224266, + "loss": 3.8428220748901367, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.019150376319885254, + "learning_rate": 0.0005041629493967257, + "loss": 3.830247402191162, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.020099453628063202, + "learning_rate": 0.0005039667427976697, + "loss": 3.838806629180908, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.021136781200766563, + "learning_rate": 0.0005037703742833391, + "loss": 3.8344273567199707, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.022857150062918663, + "learning_rate": 0.0005035738440119442, + "loss": 3.8317880630493164, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.024891739711165428, + "learning_rate": 0.0005033771521418261, + "loss": 3.837855577468872, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.02595464698970318, + "learning_rate": 0.0005031802988314557, + "loss": 3.846050262451172, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.02722409926354885, + "learning_rate": 0.0005029832842394342, + "loss": 3.8164305686950684, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.024486860260367393, + "learning_rate": 0.0005027861085244927, + "loss": 3.8099875450134277, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.021551169455051422, + "learning_rate": 0.0005025887718454922, + "loss": 3.825925827026367, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.02121506817638874, + "learning_rate": 0.0005023912743614232, + "loss": 3.870727777481079, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.01948230341076851, + "learning_rate": 0.0005021936162314058, + "loss": 3.8035314083099365, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.01896044984459877, + "learning_rate": 0.0005019957976146898, + "loss": 3.84665584564209, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.020620400086045265, + "learning_rate": 0.0005017978186706538, + "loss": 3.8338310718536377, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.020855262875556946, + "learning_rate": 0.0005015996795588062, + "loss": 3.8781003952026367, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.02153870463371277, + "learning_rate": 0.0005014013804387838, + "loss": 3.8783490657806396, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.023751024156808853, + "learning_rate": 0.0005012029214703529, + "loss": 3.846198558807373, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.023987066000699997, + "learning_rate": 0.000501004302813408, + "loss": 3.834296226501465, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.023626182228326797, + "learning_rate": 0.0005008055246279727, + "loss": 3.8591227531433105, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.025707192718982697, + "learning_rate": 0.000500606587074199, + "loss": 3.8766489028930664, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.027723029255867004, + "learning_rate": 0.0005004074903123674, + "loss": 3.88264799118042, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.0278895553201437, + "learning_rate": 0.0005002082345028864, + "loss": 3.8932995796203613, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.025622662156820297, + "learning_rate": 0.0005000088198062929, + "loss": 3.8734612464904785, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.02639615908265114, + "learning_rate": 0.0004998092463832516, + "loss": 3.786813735961914, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.026765599846839905, + "learning_rate": 0.0004996095143945554, + "loss": 3.853966474533081, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02823319099843502, + "learning_rate": 0.0004994096240011246, + "loss": 3.8321340084075928, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.023394955322146416, + "learning_rate": 0.0004992095753640072, + "loss": 3.8527331352233887, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.023874662816524506, + "learning_rate": 0.0004990093686443791, + "loss": 3.8822879791259766, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.02301311306655407, + "learning_rate": 0.000498809004003543, + "loss": 3.846186876296997, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.022630592808127403, + "learning_rate": 0.0004986084816029291, + "loss": 3.867992639541626, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.02424517087638378, + "learning_rate": 0.0004984078016040946, + "loss": 3.867293357849121, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.024116624146699905, + "learning_rate": 0.000498206964168724, + "loss": 3.846282958984375, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.02187768742442131, + "learning_rate": 0.000498005969458628, + "loss": 3.8429484367370605, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.01879809983074665, + "learning_rate": 0.0004978048176357447, + "loss": 3.863215446472168, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.020426234230399132, + "learning_rate": 0.0004976035088621385, + "loss": 3.825230121612549, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01992812193930149, + "learning_rate": 0.0004974020433000003, + "loss": 3.851665496826172, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.020693328231573105, + "learning_rate": 0.000497200421111647, + "loss": 3.906999111175537, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.021137919276952744, + "learning_rate": 0.0004969986424595221, + "loss": 3.8369035720825195, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.023285407572984695, + "learning_rate": 0.0004967967075061949, + "loss": 3.8658390045166016, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.0219135619699955, + "learning_rate": 0.0004965946164143609, + "loss": 3.873072862625122, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.020101983100175858, + "learning_rate": 0.0004963923693468413, + "loss": 3.847805976867676, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.01875176839530468, + "learning_rate": 0.0004961899664665827, + "loss": 3.7976622581481934, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01913360133767128, + "learning_rate": 0.0004959874079366576, + "loss": 3.853851318359375, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.016863619908690453, + "learning_rate": 0.0004957846939202637, + "loss": 3.877378225326538, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.017802366986870766, + "learning_rate": 0.000495581824580724, + "loss": 3.861356019973755, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.01981814205646515, + "learning_rate": 0.0004953788000814866, + "loss": 3.88613224029541, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.017508897930383682, + "learning_rate": 0.000495175620586125, + "loss": 3.8202030658721924, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.019240548834204674, + "learning_rate": 0.0004949722862583368, + "loss": 3.853827714920044, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.022305598482489586, + "learning_rate": 0.000494768797261945, + "loss": 3.885148763656616, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.02531025931239128, + "learning_rate": 0.000494565153760897, + "loss": 3.8412041664123535, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.024474849924445152, + "learning_rate": 0.0004943613559192648, + "loss": 3.865644931793213, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.020430924370884895, + "learning_rate": 0.0004941574039012444, + "loss": 3.8649442195892334, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.01906431093811989, + "learning_rate": 0.0004939532978711566, + "loss": 3.8462204933166504, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.021118057891726494, + "learning_rate": 0.0004937490379934456, + "loss": 3.8652522563934326, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.020337693393230438, + "learning_rate": 0.00049354462443268, + "loss": 3.8506221771240234, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.019090797752141953, + "learning_rate": 0.0004933400573535521, + "loss": 3.881237506866455, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.020973797887563705, + "learning_rate": 0.000493135336920878, + "loss": 3.821274757385254, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.021984362974762917, + "learning_rate": 0.0004929304632995973, + "loss": 3.8793835639953613, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.0215684212744236, + "learning_rate": 0.0004927254366547727, + "loss": 3.8259663581848145, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.021673865616321564, + "learning_rate": 0.0004925202571515908, + "loss": 3.8505711555480957, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.123871326446533, + "eval_runtime": 58.8357, + "eval_samples_per_second": 41.505, + "eval_steps_per_second": 1.309, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.0203359667211771, + "learning_rate": 0.0004923149249553608, + "loss": 3.785353660583496, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.023197868838906288, + "learning_rate": 0.0004921094402315152, + "loss": 3.7479515075683594, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.02314298041164875, + "learning_rate": 0.0004919038031456093, + "loss": 3.7697315216064453, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.0256908368319273, + "learning_rate": 0.0004916980138633213, + "loss": 3.79056978225708, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.026193121448159218, + "learning_rate": 0.0004914920725504518, + "loss": 3.7722532749176025, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.027120303362607956, + "learning_rate": 0.000491285979372924, + "loss": 3.7991981506347656, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.0266975536942482, + "learning_rate": 0.0004910797344967834, + "loss": 3.778301477432251, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.02806396596133709, + "learning_rate": 0.000490873338088198, + "loss": 3.7578704357147217, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.02893839031457901, + "learning_rate": 0.0004906667903134573, + "loss": 3.7401275634765625, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.031170370057225227, + "learning_rate": 0.0004904600913389735, + "loss": 3.780686616897583, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.03364266827702522, + "learning_rate": 0.0004902532413312799, + "loss": 3.794442653656006, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.030121006071567535, + "learning_rate": 0.0004900462404570321, + "loss": 3.779569149017334, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.03296204283833504, + "learning_rate": 0.0004898390888830067, + "loss": 3.7960336208343506, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.03381749242544174, + "learning_rate": 0.0004896317867761022, + "loss": 3.7959280014038086, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.035383883863687515, + "learning_rate": 0.000489424334303338, + "loss": 3.8008625507354736, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.03759761154651642, + "learning_rate": 0.0004892167316318548, + "loss": 3.7997028827667236, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.030651796609163284, + "learning_rate": 0.0004890089789289147, + "loss": 3.81191349029541, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.0262477844953537, + "learning_rate": 0.0004888010763618997, + "loss": 3.8046622276306152, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.030398324131965637, + "learning_rate": 0.0004885930240983136, + "loss": 3.7903637886047363, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.030515259131789207, + "learning_rate": 0.0004883848223057801, + "loss": 3.8111231327056885, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.031686048954725266, + "learning_rate": 0.00048817647115204386, + "loss": 3.781445026397705, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.025968654081225395, + "learning_rate": 0.00048796797080496936, + "loss": 3.788196563720703, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.025340190157294273, + "learning_rate": 0.00048775932143254176, + "loss": 3.763331413269043, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.024738596752285957, + "learning_rate": 0.0004875505232028661, + "loss": 3.8005969524383545, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.021990075707435608, + "learning_rate": 0.0004873415762841672, + "loss": 3.795844078063965, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.02191784791648388, + "learning_rate": 0.00048713248084478986, + "loss": 3.8062384128570557, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.023678259924054146, + "learning_rate": 0.0004869232370531986, + "loss": 3.7760679721832275, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.024137988686561584, + "learning_rate": 0.0004867138450779773, + "loss": 3.7917492389678955, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.02673223614692688, + "learning_rate": 0.0004865043050878293, + "loss": 3.8006649017333984, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.02395169623196125, + "learning_rate": 0.0004862946172515771, + "loss": 3.7902212142944336, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.02410844899713993, + "learning_rate": 0.0004860847817381625, + "loss": 3.8073291778564453, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.028109831735491753, + "learning_rate": 0.0004858747987166463, + "loss": 3.80116605758667, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.025538964197039604, + "learning_rate": 0.00048566466835620774, + "loss": 3.791151523590088, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.02393815852701664, + "learning_rate": 0.0004854543908261453, + "loss": 3.7766106128692627, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.024659741669893265, + "learning_rate": 0.0004852439662958756, + "loss": 3.789135217666626, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.02668776921927929, + "learning_rate": 0.000485033394934934, + "loss": 3.797905445098877, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.0258641317486763, + "learning_rate": 0.000484822676912974, + "loss": 3.783323287963867, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.024846022948622704, + "learning_rate": 0.0004846118123997674, + "loss": 3.7976717948913574, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.024981874972581863, + "learning_rate": 0.00048440080156520387, + "loss": 3.818732738494873, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.02556881681084633, + "learning_rate": 0.0004841896445792909, + "loss": 3.7821896076202393, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.02558571845293045, + "learning_rate": 0.000483978341612154, + "loss": 3.8143680095672607, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.02361566200852394, + "learning_rate": 0.0004837668928340362, + "loss": 3.7710189819335938, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022570716217160225, + "learning_rate": 0.0004835552984152978, + "loss": 3.775564670562744, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.022478004917502403, + "learning_rate": 0.00048334355852641664, + "loss": 3.7920689582824707, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021241728216409683, + "learning_rate": 0.0004831316733379879, + "loss": 3.7857837677001953, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.018921511247754097, + "learning_rate": 0.0004829196430207235, + "loss": 3.7659268379211426, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01926037296652794, + "learning_rate": 0.00048270746774545237, + "loss": 3.7527129650115967, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.018743667751550674, + "learning_rate": 0.0004824951476831204, + "loss": 3.767991065979004, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.017273949459195137, + "learning_rate": 0.00048228268300479015, + "loss": 3.7834174633026123, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.017054244875907898, + "learning_rate": 0.0004820700738816403, + "loss": 3.7775752544403076, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017190931364893913, + "learning_rate": 0.0004818573204849664, + "loss": 3.7709648609161377, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01665034517645836, + "learning_rate": 0.00048164442298617987, + "loss": 3.7575321197509766, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.016932843253016472, + "learning_rate": 0.00048143138155680844, + "loss": 3.7922799587249756, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01702750287950039, + "learning_rate": 0.0004812181963684958, + "loss": 3.7729780673980713, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01766599342226982, + "learning_rate": 0.00048100486759300134, + "loss": 3.7876572608947754, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01785697042942047, + "learning_rate": 0.00048079139540220024, + "loss": 3.8131699562072754, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01830688677728176, + "learning_rate": 0.0004805777799680831, + "loss": 3.790778636932373, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017348438501358032, + "learning_rate": 0.0004803640214627561, + "loss": 3.7984330654144287, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017999377101659775, + "learning_rate": 0.0004801501200584406, + "loss": 3.7756147384643555, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.01801210641860962, + "learning_rate": 0.000479936075927473, + "loss": 3.789759874343872, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.018445398658514023, + "learning_rate": 0.000479721889242305, + "loss": 3.7683000564575195, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.017544370144605637, + "learning_rate": 0.0004795075601755027, + "loss": 3.7776038646698, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.019834233447909355, + "learning_rate": 0.00047929308889974737, + "loss": 3.7559571266174316, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01791626401245594, + "learning_rate": 0.0004790784755878346, + "loss": 3.7746729850769043, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.017638562247157097, + "learning_rate": 0.00047886372041267434, + "loss": 3.7933082580566406, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.018974050879478455, + "learning_rate": 0.0004786488235472909, + "loss": 3.795713424682617, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.018694303929805756, + "learning_rate": 0.0004784337851648231, + "loss": 3.7675955295562744, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.01788092777132988, + "learning_rate": 0.0004782186054385233, + "loss": 3.786083698272705, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.01902497000992298, + "learning_rate": 0.0004780032845417579, + "loss": 3.7976233959198, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.020558880642056465, + "learning_rate": 0.000477787822648007, + "loss": 3.7500152587890625, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.020093899220228195, + "learning_rate": 0.00047757221993086443, + "loss": 3.703448534011841, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.019833799451589584, + "learning_rate": 0.00047735647656403754, + "loss": 3.756059408187866, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.093987941741943, + "eval_runtime": 57.6894, + "eval_samples_per_second": 42.33, + "eval_steps_per_second": 1.335, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.022167526185512543, + "learning_rate": 0.0004771405927213467, + "loss": 3.696444511413574, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.027874166145920753, + "learning_rate": 0.0004769245685767255, + "loss": 3.713630199432373, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.030872784554958344, + "learning_rate": 0.00047670840430422113, + "loss": 3.678295612335205, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.03246859833598137, + "learning_rate": 0.0004764921000779928, + "loss": 3.7186388969421387, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.03667069226503372, + "learning_rate": 0.0004762756560723131, + "loss": 3.7457995414733887, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.036286015063524246, + "learning_rate": 0.0004760590724615671, + "loss": 3.6881775856018066, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.03413614630699158, + "learning_rate": 0.0004758423494202522, + "loss": 3.6999263763427734, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.03384552150964737, + "learning_rate": 0.0004756254871229782, + "loss": 3.6959803104400635, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.03227181360125542, + "learning_rate": 0.00047540848574446724, + "loss": 3.7182722091674805, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.036366235464811325, + "learning_rate": 0.0004751913454595535, + "loss": 3.7116851806640625, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.03573326766490936, + "learning_rate": 0.0004749740664431828, + "loss": 3.741865873336792, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.03482096269726753, + "learning_rate": 0.000474756648870413, + "loss": 3.7195520401000977, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.03232625126838684, + "learning_rate": 0.0004745390929164134, + "loss": 3.7443251609802246, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.03421664610505104, + "learning_rate": 0.0004743213987564651, + "loss": 3.6867361068725586, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.03585013002157211, + "learning_rate": 0.0004741035665659603, + "loss": 3.7492403984069824, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.029985329136252403, + "learning_rate": 0.00047388559652040236, + "loss": 3.705209493637085, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.02729589305818081, + "learning_rate": 0.0004736674887954059, + "loss": 3.711106538772583, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.029642216861248016, + "learning_rate": 0.0004734492435666963, + "loss": 3.7274727821350098, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.02816903032362461, + "learning_rate": 0.0004732308610101099, + "loss": 3.7181742191314697, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.027143625542521477, + "learning_rate": 0.00047301234130159346, + "loss": 3.719235897064209, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.025802744552493095, + "learning_rate": 0.00047279368461720445, + "loss": 3.771820545196533, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.026424532756209373, + "learning_rate": 0.0004725748911331106, + "loss": 3.7516651153564453, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.029677648097276688, + "learning_rate": 0.0004723559610255898, + "loss": 3.7058472633361816, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.031548649072647095, + "learning_rate": 0.0004721368944710302, + "loss": 3.756992816925049, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.027667395770549774, + "learning_rate": 0.00047191769164592974, + "loss": 3.742711067199707, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.027639545500278473, + "learning_rate": 0.000471698352726896, + "loss": 3.728883743286133, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.02950332500040531, + "learning_rate": 0.0004714788778906467, + "loss": 3.7160212993621826, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.026302607730031013, + "learning_rate": 0.0004712592673140084, + "loss": 3.735713481903076, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.022483112290501595, + "learning_rate": 0.00047103952117391764, + "loss": 3.702686309814453, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.021729422733187675, + "learning_rate": 0.00047081963964741986, + "loss": 3.7195816040039062, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.021159254014492035, + "learning_rate": 0.0004705996229116696, + "loss": 3.684321880340576, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.02233424223959446, + "learning_rate": 0.0004703794711439304, + "loss": 3.7233057022094727, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.02142048068344593, + "learning_rate": 0.0004701591845215744, + "loss": 3.7083680629730225, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.022331461310386658, + "learning_rate": 0.00046993876322208274, + "loss": 3.7408461570739746, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.020629866048693657, + "learning_rate": 0.0004697182074230448, + "loss": 3.759129047393799, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.02174685336649418, + "learning_rate": 0.0004694975173021586, + "loss": 3.72351336479187, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.021538110449910164, + "learning_rate": 0.0004692766930372299, + "loss": 3.716416597366333, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.020800625905394554, + "learning_rate": 0.00046905573480617285, + "loss": 3.726954936981201, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.02039588987827301, + "learning_rate": 0.00046883464278700966, + "loss": 3.7336390018463135, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.02107219770550728, + "learning_rate": 0.0004686134171578702, + "loss": 3.722311496734619, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.02068956382572651, + "learning_rate": 0.00046839205809699183, + "loss": 3.767139434814453, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.020190054550766945, + "learning_rate": 0.0004681705657827195, + "loss": 3.6960511207580566, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.018814411014318466, + "learning_rate": 0.0004679489403935058, + "loss": 3.70375919342041, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.017568519338965416, + "learning_rate": 0.00046772718210791005, + "loss": 3.756627321243286, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.018188195303082466, + "learning_rate": 0.00046750529110459903, + "loss": 3.6964683532714844, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.018925746902823448, + "learning_rate": 0.0004672832675623463, + "loss": 3.731172561645508, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.01690756157040596, + "learning_rate": 0.0004670611116600322, + "loss": 3.6955318450927734, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016300756484270096, + "learning_rate": 0.00046683882357664375, + "loss": 3.762632369995117, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.016405068337917328, + "learning_rate": 0.0004666164034912744, + "loss": 3.75130558013916, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.017687877640128136, + "learning_rate": 0.00046639385158312406, + "loss": 3.687530755996704, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.017524344846606255, + "learning_rate": 0.00046617116803149874, + "loss": 3.738790273666382, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.01714804768562317, + "learning_rate": 0.00046594835301581053, + "loss": 3.69827938079834, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018042324110865593, + "learning_rate": 0.00046572540671557745, + "loss": 3.71600341796875, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018648479133844376, + "learning_rate": 0.00046550232931042346, + "loss": 3.742591619491577, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.019198937341570854, + "learning_rate": 0.00046527912098007787, + "loss": 3.739820957183838, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.01796373911201954, + "learning_rate": 0.00046505578190437564, + "loss": 3.719025135040283, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01836620457470417, + "learning_rate": 0.0004648323122632569, + "loss": 3.741434097290039, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.02192627638578415, + "learning_rate": 0.0004646087122367673, + "loss": 3.7553882598876953, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.02147185243666172, + "learning_rate": 0.00046438498200505727, + "loss": 3.7705790996551514, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01855960488319397, + "learning_rate": 0.0004641611217483821, + "loss": 3.704925537109375, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.01925475336611271, + "learning_rate": 0.00046393713164710217, + "loss": 3.711491346359253, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019796503707766533, + "learning_rate": 0.00046371301188168204, + "loss": 3.7241992950439453, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.018511971458792686, + "learning_rate": 0.0004634887626326911, + "loss": 3.6785435676574707, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.018214019015431404, + "learning_rate": 0.00046326438408080293, + "loss": 3.7299962043762207, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.018912190571427345, + "learning_rate": 0.00046303987640679517, + "loss": 3.773698329925537, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.01939854398369789, + "learning_rate": 0.0004628152397915498, + "loss": 3.682321071624756, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.017859861254692078, + "learning_rate": 0.00046259047441605215, + "loss": 3.7254951000213623, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.019249707460403442, + "learning_rate": 0.0004623655804613919, + "loss": 3.7263433933258057, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.019221967086195946, + "learning_rate": 0.00046214055810876194, + "loss": 3.7229225635528564, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01871953159570694, + "learning_rate": 0.00046191540753945886, + "loss": 3.7248311042785645, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01747182011604309, + "learning_rate": 0.00046169012893488204, + "loss": 3.699913740158081, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.019630730152130127, + "learning_rate": 0.00046146472247653484, + "loss": 3.787783145904541, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.072091102600098, + "eval_runtime": 63.1589, + "eval_samples_per_second": 38.664, + "eval_steps_per_second": 1.219, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.022918934002518654, + "learning_rate": 0.0004612391883460228, + "loss": 3.666675567626953, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.0264838095754385, + "learning_rate": 0.00046101352672505493, + "loss": 3.6371259689331055, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.028463594615459442, + "learning_rate": 0.0004607877377954426, + "loss": 3.657130241394043, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.02900792844593525, + "learning_rate": 0.00046056182173909995, + "loss": 3.6365628242492676, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.028593607246875763, + "learning_rate": 0.0004603357787380435, + "loss": 3.647258996963501, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.02878790907561779, + "learning_rate": 0.0004601096089743919, + "loss": 3.682035446166992, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.02980183996260166, + "learning_rate": 0.0004598833126303661, + "loss": 3.623297691345215, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.03375725448131561, + "learning_rate": 0.00045965688988828884, + "loss": 3.6240830421447754, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.035045325756073, + "learning_rate": 0.00045943034093058507, + "loss": 3.6372249126434326, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.03519923985004425, + "learning_rate": 0.000459203665939781, + "loss": 3.6157641410827637, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.03468473628163338, + "learning_rate": 0.00045897686509850487, + "loss": 3.6553335189819336, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.034817393869161606, + "learning_rate": 0.0004587499385894856, + "loss": 3.653806686401367, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.03549059107899666, + "learning_rate": 0.000458522886595554, + "loss": 3.6730189323425293, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.028001410886645317, + "learning_rate": 0.0004582957092996418, + "loss": 3.662490129470825, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.02532520331442356, + "learning_rate": 0.0004580684068847817, + "loss": 3.629936456680298, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.028215918689966202, + "learning_rate": 0.0004578409795341069, + "loss": 3.6507444381713867, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.026823483407497406, + "learning_rate": 0.0004576134274308517, + "loss": 3.661811113357544, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.0281058382242918, + "learning_rate": 0.0004573857507583506, + "loss": 3.6805872917175293, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.02548503689467907, + "learning_rate": 0.00045715794970003874, + "loss": 3.6586413383483887, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.024795135483145714, + "learning_rate": 0.0004569300244394513, + "loss": 3.654989004135132, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.025501688942313194, + "learning_rate": 0.0004567019751602234, + "loss": 3.6915903091430664, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.02377370186150074, + "learning_rate": 0.00045647380204609034, + "loss": 3.660588264465332, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.022150425240397453, + "learning_rate": 0.00045624550528088704, + "loss": 3.64870023727417, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.02415974996984005, + "learning_rate": 0.00045601708504854803, + "loss": 3.6498191356658936, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.02140078693628311, + "learning_rate": 0.0004557885415331074, + "loss": 3.703489065170288, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.022121304646134377, + "learning_rate": 0.00045555987491869865, + "loss": 3.6606979370117188, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.02168990671634674, + "learning_rate": 0.00045533108538955416, + "loss": 3.6801178455352783, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.020228713750839233, + "learning_rate": 0.0004551021731300055, + "loss": 3.6707205772399902, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.018500352278351784, + "learning_rate": 0.0004548731383244834, + "loss": 3.647545576095581, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.020223816856741905, + "learning_rate": 0.0004546439811575169, + "loss": 3.644932985305786, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.019502121955156326, + "learning_rate": 0.00045441470181373375, + "loss": 3.62888240814209, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.01926770992577076, + "learning_rate": 0.0004541853004778603, + "loss": 3.675457715988159, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01888471283018589, + "learning_rate": 0.00045395577733472087, + "loss": 3.636434316635132, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.0201494712382555, + "learning_rate": 0.0004537261325692383, + "loss": 3.679147720336914, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.02178201824426651, + "learning_rate": 0.0004534963663664332, + "loss": 3.671905040740967, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.024073559790849686, + "learning_rate": 0.00045326647891142394, + "loss": 3.6701040267944336, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.02135559171438217, + "learning_rate": 0.0004530364703894268, + "loss": 3.648519992828369, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.019264448434114456, + "learning_rate": 0.0004528063409857554, + "loss": 3.6985023021698, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.01922295056283474, + "learning_rate": 0.0004525760908858209, + "loss": 3.651756763458252, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.01978331431746483, + "learning_rate": 0.00045234572027513174, + "loss": 3.679884910583496, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.01968599483370781, + "learning_rate": 0.00045211522933929317, + "loss": 3.6571812629699707, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.020741639658808708, + "learning_rate": 0.0004518846182640077, + "loss": 3.6505298614501953, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.01842017099261284, + "learning_rate": 0.0004516538872350744, + "loss": 3.6904940605163574, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.018004486337304115, + "learning_rate": 0.0004514230364383893, + "loss": 3.7282047271728516, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.018821008503437042, + "learning_rate": 0.0004511920660599444, + "loss": 3.668164014816284, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.018929854035377502, + "learning_rate": 0.0004509609762858287, + "loss": 3.6672980785369873, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.01979168877005577, + "learning_rate": 0.0004507297673022269, + "loss": 3.6684999465942383, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.01774384267628193, + "learning_rate": 0.0004504984392954199, + "loss": 3.6577749252319336, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01876000687479973, + "learning_rate": 0.00045026699245178444, + "loss": 3.6629252433776855, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.018197249621152878, + "learning_rate": 0.0004500354269577932, + "loss": 3.688983678817749, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.019496265798807144, + "learning_rate": 0.00044980374300001414, + "loss": 3.6602022647857666, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01825535111129284, + "learning_rate": 0.00044957194076511095, + "loss": 3.6889238357543945, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.017113137990236282, + "learning_rate": 0.0004493400204398426, + "loss": 3.684070587158203, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.017993126064538956, + "learning_rate": 0.00044910798221106283, + "loss": 3.675426721572876, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.018354708328843117, + "learning_rate": 0.00044887582626572105, + "loss": 3.675588607788086, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.016362832859158516, + "learning_rate": 0.0004486435527908606, + "loss": 3.6709303855895996, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01772182621061802, + "learning_rate": 0.0004484111619736205, + "loss": 3.6884102821350098, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.019271764904260635, + "learning_rate": 0.0004481786540012336, + "loss": 3.6468968391418457, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.019040528684854507, + "learning_rate": 0.00044794602906102747, + "loss": 3.690387010574341, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.019022610038518906, + "learning_rate": 0.00044771328734042367, + "loss": 3.689873218536377, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.018533391878008842, + "learning_rate": 0.00044748042902693815, + "loss": 3.6877284049987793, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.020825615152716637, + "learning_rate": 0.0004472474543081805, + "loss": 3.662257671356201, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.021880341693758965, + "learning_rate": 0.00044701436337185444, + "loss": 3.650700092315674, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.019974704831838608, + "learning_rate": 0.00044678115640575704, + "loss": 3.684177875518799, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.01903127320110798, + "learning_rate": 0.0004465478335977788, + "loss": 3.658148765563965, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.020999105647206306, + "learning_rate": 0.00044631439513590393, + "loss": 3.68747615814209, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.02138398587703705, + "learning_rate": 0.00044608084120820933, + "loss": 3.656355381011963, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.02216442860662937, + "learning_rate": 0.0004458471720028655, + "loss": 3.6860342025756836, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.019359121099114418, + "learning_rate": 0.0004456133877081352, + "loss": 3.654944658279419, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.020984770730137825, + "learning_rate": 0.0004453794885123744, + "loss": 3.6885194778442383, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.02423972077667713, + "learning_rate": 0.00044514547460403135, + "loss": 3.672712802886963, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.024544453248381615, + "learning_rate": 0.00044491134617164694, + "loss": 3.6484768390655518, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.060352325439453, + "eval_runtime": 57.9556, + "eval_samples_per_second": 42.136, + "eval_steps_per_second": 1.329, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.024072550237178802, + "learning_rate": 0.000444677103403854, + "loss": 3.5796468257904053, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.028701433911919594, + "learning_rate": 0.0004444427464893779, + "loss": 3.604318618774414, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.028119578957557678, + "learning_rate": 0.00044420827561703573, + "loss": 3.608759641647339, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.025335192680358887, + "learning_rate": 0.0004439736909757365, + "loss": 3.5760319232940674, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.0273929201066494, + "learning_rate": 0.0004437389927544805, + "loss": 3.559441089630127, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.026567714288830757, + "learning_rate": 0.0004435041811423603, + "loss": 3.626889228820801, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.027587885037064552, + "learning_rate": 0.0004432692563285591, + "loss": 3.6226272583007812, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.02935059368610382, + "learning_rate": 0.0004430342185023517, + "loss": 3.601921558380127, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.02982175536453724, + "learning_rate": 0.0004427990678531038, + "loss": 3.59737491607666, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.028570258989930153, + "learning_rate": 0.0004425638045702719, + "loss": 3.56337833404541, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.029890350997447968, + "learning_rate": 0.0004423284288434035, + "loss": 3.6294307708740234, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.03206814453005791, + "learning_rate": 0.00044209294086213657, + "loss": 3.590505599975586, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.03169345110654831, + "learning_rate": 0.00044185734081619957, + "loss": 3.581584930419922, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.03313758969306946, + "learning_rate": 0.000441621628895411, + "loss": 3.6088438034057617, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.03317603841423988, + "learning_rate": 0.00044138580528967986, + "loss": 3.6340456008911133, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.034738484770059586, + "learning_rate": 0.0004411498701890051, + "loss": 3.5814096927642822, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.036428093910217285, + "learning_rate": 0.0004409138237834751, + "loss": 3.5939526557922363, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.03293406963348389, + "learning_rate": 0.00044067766626326835, + "loss": 3.577897787094116, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.031030720099806786, + "learning_rate": 0.00044044139781865255, + "loss": 3.6077990531921387, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.03441236540675163, + "learning_rate": 0.00044020501863998516, + "loss": 3.591923713684082, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.03500965237617493, + "learning_rate": 0.00043996852891771236, + "loss": 3.6171927452087402, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.034443050622940063, + "learning_rate": 0.00043973192884236983, + "loss": 3.619724750518799, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.03435051813721657, + "learning_rate": 0.0004394952186045819, + "loss": 3.6443169116973877, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.03230702877044678, + "learning_rate": 0.00043925839839506167, + "loss": 3.5912671089172363, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.03232988342642784, + "learning_rate": 0.0004390214684046111, + "loss": 3.618985414505005, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.029298555105924606, + "learning_rate": 0.00043878442882412015, + "loss": 3.6166677474975586, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.026723450049757957, + "learning_rate": 0.00043854727984456746, + "loss": 3.61362886428833, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.028280874714255333, + "learning_rate": 0.0004383100216570197, + "loss": 3.6410951614379883, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.026700228452682495, + "learning_rate": 0.0004380726544526316, + "loss": 3.5804049968719482, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.023092713207006454, + "learning_rate": 0.00043783517842264534, + "loss": 3.595339298248291, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.024123897776007652, + "learning_rate": 0.0004375975937583914, + "loss": 3.647569417953491, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.02505572699010372, + "learning_rate": 0.0004373599006512871, + "loss": 3.6276652812957764, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.024523509666323662, + "learning_rate": 0.00043712209929283786, + "loss": 3.59525728225708, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.024351581931114197, + "learning_rate": 0.00043688418987463567, + "loss": 3.6081674098968506, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.02530614472925663, + "learning_rate": 0.00043664617258835995, + "loss": 3.6632542610168457, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.023438062518835068, + "learning_rate": 0.0004364080476257769, + "loss": 3.6345431804656982, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.021853657439351082, + "learning_rate": 0.0004361698151787396, + "loss": 3.598644256591797, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.022487422451376915, + "learning_rate": 0.0004359314754391874, + "loss": 3.5966577529907227, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.021353935822844505, + "learning_rate": 0.0004356930285991466, + "loss": 3.6328375339508057, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.021180452778935432, + "learning_rate": 0.00043545447485072926, + "loss": 3.6285812854766846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.020491832867264748, + "learning_rate": 0.000435215814386134, + "loss": 3.592944622039795, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.021603770554065704, + "learning_rate": 0.0004349770473976453, + "loss": 3.613507032394409, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.02146158367395401, + "learning_rate": 0.0004347381740776332, + "loss": 3.6409506797790527, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.02102433145046234, + "learning_rate": 0.0004344991946185538, + "loss": 3.6179330348968506, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.018684551119804382, + "learning_rate": 0.00043426010921294837, + "loss": 3.6385130882263184, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.018416227772831917, + "learning_rate": 0.0004340209180534438, + "loss": 3.642151355743408, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.01915144734084606, + "learning_rate": 0.000433781621332752, + "loss": 3.6089656352996826, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.018299665302038193, + "learning_rate": 0.0004335422192436703, + "loss": 3.6063547134399414, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.018962757661938667, + "learning_rate": 0.00043330271197908023, + "loss": 3.625917434692383, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.017437821254134178, + "learning_rate": 0.00043306309973194874, + "loss": 3.600015640258789, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.017913147807121277, + "learning_rate": 0.000432823382695327, + "loss": 3.6428744792938232, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.018343618139624596, + "learning_rate": 0.0004325835610623508, + "loss": 3.635901927947998, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.020372850820422173, + "learning_rate": 0.00043234363502623994, + "loss": 3.620181083679199, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.02037600241601467, + "learning_rate": 0.0004321036047802985, + "loss": 3.61129093170166, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.021827151998877525, + "learning_rate": 0.00043186347051791466, + "loss": 3.5999555587768555, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.021806657314300537, + "learning_rate": 0.0004316232324325602, + "loss": 3.642509937286377, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02050699107348919, + "learning_rate": 0.0004313828907177906, + "loss": 3.597583293914795, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.020436758175492287, + "learning_rate": 0.0004311424455672448, + "loss": 3.6079912185668945, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.019341666251420975, + "learning_rate": 0.0004309018971746451, + "loss": 3.6196017265319824, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.022348817437887192, + "learning_rate": 0.000430661245733797, + "loss": 3.619025468826294, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.019361121580004692, + "learning_rate": 0.00043042049143858914, + "loss": 3.657182455062866, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.019258935004472733, + "learning_rate": 0.0004301796344829928, + "loss": 3.6408963203430176, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.020622272044420242, + "learning_rate": 0.00042993867506106205, + "loss": 3.6247220039367676, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.019510755315423012, + "learning_rate": 0.0004296976133669336, + "loss": 3.6264584064483643, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.02078665979206562, + "learning_rate": 0.0004294564495948266, + "loss": 3.6364388465881348, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01791800744831562, + "learning_rate": 0.0004292151839390421, + "loss": 3.6029374599456787, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.01922667771577835, + "learning_rate": 0.0004289738165939636, + "loss": 3.582937717437744, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.018473893404006958, + "learning_rate": 0.00042873234775405654, + "loss": 3.6141819953918457, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01846298575401306, + "learning_rate": 0.00042849077761386784, + "loss": 3.638786554336548, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.01876973733305931, + "learning_rate": 0.0004282491063680263, + "loss": 3.617130756378174, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.017995499074459076, + "learning_rate": 0.00042800733421124207, + "loss": 3.645305633544922, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.02105429396033287, + "learning_rate": 0.00042776546133830646, + "loss": 3.6444876194000244, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.049062252044678, + "eval_runtime": 58.0113, + "eval_samples_per_second": 42.095, + "eval_steps_per_second": 1.327, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.02128119394183159, + "learning_rate": 0.00042752348794409227, + "loss": 3.524954080581665, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.028249500319361687, + "learning_rate": 0.00042728141422355287, + "loss": 3.5311479568481445, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.03145391866564751, + "learning_rate": 0.0004270392403717229, + "loss": 3.559330940246582, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.02994118444621563, + "learning_rate": 0.00042679696658371743, + "loss": 3.5257620811462402, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.027448246255517006, + "learning_rate": 0.00042655459305473195, + "loss": 3.5242764949798584, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.0296134352684021, + "learning_rate": 0.0004263121199800425, + "loss": 3.5317091941833496, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.029576804488897324, + "learning_rate": 0.0004260695475550054, + "loss": 3.5340428352355957, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.02941269613802433, + "learning_rate": 0.0004258268759750566, + "loss": 3.5372023582458496, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.03126693144440651, + "learning_rate": 0.0004255841054357124, + "loss": 3.5437135696411133, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.03335367515683174, + "learning_rate": 0.0004253412361325686, + "loss": 3.5598273277282715, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.03028946928679943, + "learning_rate": 0.0004250982682613006, + "loss": 3.5190839767456055, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.028629133477807045, + "learning_rate": 0.00042485520201766315, + "loss": 3.5138821601867676, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.027070026844739914, + "learning_rate": 0.0004246120375974905, + "loss": 3.530958414077759, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.028837760910391808, + "learning_rate": 0.0004243687751966957, + "loss": 3.5233962535858154, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.029884297400712967, + "learning_rate": 0.0004241254150112708, + "loss": 3.5639758110046387, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.024223247542977333, + "learning_rate": 0.000423881957237287, + "loss": 3.5555152893066406, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.026992836967110634, + "learning_rate": 0.0004236384020708935, + "loss": 3.540196418762207, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.02818330004811287, + "learning_rate": 0.0004233947497083185, + "loss": 3.5546071529388428, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.026401042938232422, + "learning_rate": 0.0004231510003458682, + "loss": 3.5556247234344482, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.026020489633083344, + "learning_rate": 0.0004229071541799272, + "loss": 3.568004608154297, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.02803782932460308, + "learning_rate": 0.0004226632114069578, + "loss": 3.5828235149383545, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.02737978659570217, + "learning_rate": 0.0004224191722235004, + "loss": 3.5642499923706055, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.02406371757388115, + "learning_rate": 0.00042217503682617286, + "loss": 3.5530993938446045, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.02230692468583584, + "learning_rate": 0.0004219308054116706, + "loss": 3.5573501586914062, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.021352406591176987, + "learning_rate": 0.0004216864781767667, + "loss": 3.5526974201202393, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.02227999083697796, + "learning_rate": 0.00042144205531831103, + "loss": 3.5866928100585938, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.024190247058868408, + "learning_rate": 0.0004211975370332308, + "loss": 3.517117977142334, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.023196619004011154, + "learning_rate": 0.00042095292351852976, + "loss": 3.4826502799987793, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.02388714998960495, + "learning_rate": 0.0004207082149712888, + "loss": 3.5588626861572266, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.024692339822649956, + "learning_rate": 0.0004204634115886651, + "loss": 3.54828143119812, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.023078206926584244, + "learning_rate": 0.0004202185135678924, + "loss": 3.5679030418395996, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.020665735006332397, + "learning_rate": 0.00041997352110628033, + "loss": 3.5500283241271973, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.02022322081029415, + "learning_rate": 0.00041972843440121527, + "loss": 3.592104911804199, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.021137656643986702, + "learning_rate": 0.00041948325365015894, + "loss": 3.583833694458008, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.021026931703090668, + "learning_rate": 0.0004192379790506491, + "loss": 3.556403875350952, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.020423648878932, + "learning_rate": 0.00041899261080029904, + "loss": 3.5839173793792725, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.020935002714395523, + "learning_rate": 0.00041874714909679754, + "loss": 3.5794217586517334, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.020799759775400162, + "learning_rate": 0.00041850159413790863, + "loss": 3.6033859252929688, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.019480884075164795, + "learning_rate": 0.0004182559461214715, + "loss": 3.580097198486328, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.019300777465105057, + "learning_rate": 0.0004180102052454004, + "loss": 3.550474166870117, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.020265772938728333, + "learning_rate": 0.00041776437170768427, + "loss": 3.5711655616760254, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.019972780719399452, + "learning_rate": 0.0004175184457063868, + "loss": 3.5636885166168213, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.019635623320937157, + "learning_rate": 0.000417272427439646, + "loss": 3.5851311683654785, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.021451901644468307, + "learning_rate": 0.00041702631710567443, + "loss": 3.5956759452819824, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.019456414505839348, + "learning_rate": 0.00041678011490275875, + "loss": 3.595731258392334, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.019757689908146858, + "learning_rate": 0.00041653382102925957, + "loss": 3.569652557373047, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.020571604371070862, + "learning_rate": 0.00041628743568361147, + "loss": 3.6007747650146484, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.02153848111629486, + "learning_rate": 0.00041604095906432265, + "loss": 3.5915260314941406, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.020875850692391396, + "learning_rate": 0.00041579439136997493, + "loss": 3.5582690238952637, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.02089419960975647, + "learning_rate": 0.00041554773279922337, + "loss": 3.5756618976593018, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01961098238825798, + "learning_rate": 0.0004153009835507963, + "loss": 3.5918776988983154, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.021075716242194176, + "learning_rate": 0.0004150541438234952, + "loss": 3.5342202186584473, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.020745476707816124, + "learning_rate": 0.00041480721381619435, + "loss": 3.5976028442382812, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.02001352794468403, + "learning_rate": 0.00041456019372784086, + "loss": 3.5686182975769043, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.021684160456061363, + "learning_rate": 0.0004143130837574543, + "loss": 3.553445339202881, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.022551920264959335, + "learning_rate": 0.0004140658841041267, + "loss": 3.5712709426879883, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.022758224979043007, + "learning_rate": 0.0004138185949670224, + "loss": 3.6101667881011963, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.020714636892080307, + "learning_rate": 0.00041357121654537766, + "loss": 3.5740506649017334, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.022694187238812447, + "learning_rate": 0.00041332374903850104, + "loss": 3.5663979053497314, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.021664340049028397, + "learning_rate": 0.00041307619264577234, + "loss": 3.5830888748168945, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.019072270020842552, + "learning_rate": 0.0004128285475666436, + "loss": 3.5861973762512207, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.020153764635324478, + "learning_rate": 0.0004125808140006378, + "loss": 3.566444158554077, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019666558131575584, + "learning_rate": 0.0004123329921473494, + "loss": 3.5789592266082764, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.02101624570786953, + "learning_rate": 0.0004120850822064439, + "loss": 3.551722526550293, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.01951020397245884, + "learning_rate": 0.00041183708437765815, + "loss": 3.6069014072418213, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.01985476352274418, + "learning_rate": 0.00041158899886079925, + "loss": 3.5944924354553223, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02195335365831852, + "learning_rate": 0.00041134082585574534, + "loss": 3.5828046798706055, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02105172537267208, + "learning_rate": 0.00041109256556244497, + "loss": 3.5830953121185303, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.01888827420771122, + "learning_rate": 0.00041084421818091685, + "loss": 3.5656847953796387, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.017867228016257286, + "learning_rate": 0.0004105957839112501, + "loss": 3.5792243480682373, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.018577905371785164, + "learning_rate": 0.0004103472629536036, + "loss": 3.56992506980896, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.020045632496476173, + "learning_rate": 0.0004100986555082063, + "loss": 3.590876579284668, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.044339656829834, + "eval_runtime": 57.1435, + "eval_samples_per_second": 42.735, + "eval_steps_per_second": 1.347, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.024135645478963852, + "learning_rate": 0.0004098499617753567, + "loss": 3.4856467247009277, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.031053155660629272, + "learning_rate": 0.00040960118195542273, + "loss": 3.455773115158081, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.035190410912036896, + "learning_rate": 0.00040935231624884204, + "loss": 3.506619930267334, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.03611094877123833, + "learning_rate": 0.0004091033648561211, + "loss": 3.5227696895599365, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.034121073782444, + "learning_rate": 0.00040885432797783543, + "loss": 3.4970145225524902, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.03629382699728012, + "learning_rate": 0.0004086052058146297, + "loss": 3.506425619125366, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.03709045797586441, + "learning_rate": 0.00040835599856721725, + "loss": 3.499274969100952, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.034826792776584625, + "learning_rate": 0.00040810670643637954, + "loss": 3.5029163360595703, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.03344690054655075, + "learning_rate": 0.000407857329622967, + "loss": 3.4816911220550537, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.03136949986219406, + "learning_rate": 0.00040760786832789786, + "loss": 3.5154659748077393, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.030533097684383392, + "learning_rate": 0.0004073583227521588, + "loss": 3.509666919708252, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.03298318013548851, + "learning_rate": 0.0004071086930968039, + "loss": 3.5096616744995117, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.031186232343316078, + "learning_rate": 0.00040685897956295545, + "loss": 3.4752590656280518, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.028008371591567993, + "learning_rate": 0.0004066091823518031, + "loss": 3.498826503753662, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.028206519782543182, + "learning_rate": 0.00040635930166460385, + "loss": 3.5281882286071777, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.024417119100689888, + "learning_rate": 0.00040610933770268226, + "loss": 3.506791830062866, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.025321070104837418, + "learning_rate": 0.00040585929066742964, + "loss": 3.494256019592285, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.023430505767464638, + "learning_rate": 0.00040560916076030435, + "loss": 3.533010482788086, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.023371374234557152, + "learning_rate": 0.00040535894818283154, + "loss": 3.5197086334228516, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.02414027415215969, + "learning_rate": 0.0004051086531366031, + "loss": 3.4707889556884766, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.026652060449123383, + "learning_rate": 0.000404858275823277, + "loss": 3.498159885406494, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.024486590176820755, + "learning_rate": 0.000404607816444578, + "loss": 3.50034236907959, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.023153992369771004, + "learning_rate": 0.00040435727520229644, + "loss": 3.499504804611206, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.02222427912056446, + "learning_rate": 0.0004041066522982891, + "loss": 3.4953579902648926, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.023033641278743744, + "learning_rate": 0.000403855947934478, + "loss": 3.5181496143341064, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.02141181193292141, + "learning_rate": 0.00040360516231285155, + "loss": 3.552860736846924, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.023448696359992027, + "learning_rate": 0.00040335429563546286, + "loss": 3.5153324604034424, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.022410666570067406, + "learning_rate": 0.0004031033481044308, + "loss": 3.4963512420654297, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.020326802507042885, + "learning_rate": 0.00040285231992193924, + "loss": 3.528319835662842, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021117394790053368, + "learning_rate": 0.0004026012112902372, + "loss": 3.5192837715148926, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.021317072212696075, + "learning_rate": 0.0004023500224116381, + "loss": 3.5081021785736084, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.021116144955158234, + "learning_rate": 0.00040209875348852037, + "loss": 3.5167174339294434, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019834019243717194, + "learning_rate": 0.00040184740472332705, + "loss": 3.511960983276367, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.02089378982782364, + "learning_rate": 0.000401595976318565, + "loss": 3.493330955505371, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.021129710599780083, + "learning_rate": 0.00040134446847680584, + "loss": 3.544156551361084, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.021163299679756165, + "learning_rate": 0.0004010928814006846, + "loss": 3.5261003971099854, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.02105306088924408, + "learning_rate": 0.0004008412152929007, + "loss": 3.543973445892334, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.02150859497487545, + "learning_rate": 0.0004005894703562168, + "loss": 3.487217426300049, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.021666452288627625, + "learning_rate": 0.0004003376467934593, + "loss": 3.4933671951293945, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.02180723287165165, + "learning_rate": 0.0004000857448075178, + "loss": 3.55198073387146, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.02041625790297985, + "learning_rate": 0.0003998337646013455, + "loss": 3.529571533203125, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.022977063432335854, + "learning_rate": 0.00039958170637795806, + "loss": 3.526432752609253, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.022725161164999008, + "learning_rate": 0.0003993295703404341, + "loss": 3.5258829593658447, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.021337240934371948, + "learning_rate": 0.00039907735669191544, + "loss": 3.4897351264953613, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.020783131942152977, + "learning_rate": 0.00039882506563560573, + "loss": 3.5497589111328125, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.022259317338466644, + "learning_rate": 0.0003985726973747715, + "loss": 3.5225064754486084, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.020283272489905357, + "learning_rate": 0.00039832025211274113, + "loss": 3.5453078746795654, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.020656265318393707, + "learning_rate": 0.00039806773005290544, + "loss": 3.533703327178955, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.02104976586997509, + "learning_rate": 0.0003978151313987168, + "loss": 3.4885146617889404, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.021411150693893433, + "learning_rate": 0.00039756245635368933, + "loss": 3.533268690109253, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.021906673908233643, + "learning_rate": 0.00039730970512139876, + "loss": 3.5483224391937256, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.02415728010237217, + "learning_rate": 0.0003970568779054822, + "loss": 3.5084290504455566, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.023435061797499657, + "learning_rate": 0.000396803974909638, + "loss": 3.5498428344726562, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.02183385379612446, + "learning_rate": 0.0003965509963376255, + "loss": 3.484086751937866, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.022856134921312332, + "learning_rate": 0.000396297942393265, + "loss": 3.5339231491088867, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.021679896861314774, + "learning_rate": 0.0003960448132804375, + "loss": 3.547191619873047, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.021262124180793762, + "learning_rate": 0.0003957916092030845, + "loss": 3.5263822078704834, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.02265402302145958, + "learning_rate": 0.00039553833036520803, + "loss": 3.5361135005950928, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.020080721005797386, + "learning_rate": 0.0003952849769708702, + "loss": 3.5247459411621094, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.02187187783420086, + "learning_rate": 0.0003950315492241932, + "loss": 3.54129695892334, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.02417500875890255, + "learning_rate": 0.0003947780473293593, + "loss": 3.598796844482422, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.02416270412504673, + "learning_rate": 0.00039452447149061054, + "loss": 3.5207936763763428, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.021958643570542336, + "learning_rate": 0.0003942708219122481, + "loss": 3.5395452976226807, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.021813420578837395, + "learning_rate": 0.000394017098798633, + "loss": 3.5469255447387695, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020607495680451393, + "learning_rate": 0.0003937633023541854, + "loss": 3.554851531982422, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.02091527171432972, + "learning_rate": 0.0003935094327833845, + "loss": 3.492576837539673, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.021089257672429085, + "learning_rate": 0.0003932554902907683, + "loss": 3.5154709815979004, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.019659532234072685, + "learning_rate": 0.0003930014750809338, + "loss": 3.5444283485412598, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01896163448691368, + "learning_rate": 0.0003927473873585365, + "loss": 3.589625358581543, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020288823172450066, + "learning_rate": 0.0003924932273282903, + "loss": 3.522411823272705, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.020222362130880356, + "learning_rate": 0.00039223899519496723, + "loss": 3.5244359970092773, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.022349944338202477, + "learning_rate": 0.00039198469116339757, + "loss": 3.5362396240234375, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.047028064727783, + "eval_runtime": 62.8948, + "eval_samples_per_second": 38.827, + "eval_steps_per_second": 1.224, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.02202913723886013, + "learning_rate": 0.0003917303154384694, + "loss": 3.4249138832092285, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.03153969347476959, + "learning_rate": 0.00039147586822512885, + "loss": 3.4436073303222656, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.03747870773077011, + "learning_rate": 0.0003912213497283793, + "loss": 3.4879579544067383, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.036144088953733444, + "learning_rate": 0.0003909667601532819, + "loss": 3.419151782989502, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.03983336314558983, + "learning_rate": 0.00039071209970495445, + "loss": 3.4331436157226562, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.045912645757198334, + "learning_rate": 0.00039045736858857273, + "loss": 3.4304769039154053, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.04005753993988037, + "learning_rate": 0.0003902025670093687, + "loss": 3.478600025177002, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.03488151356577873, + "learning_rate": 0.00038994769517263156, + "loss": 3.4650232791900635, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.03603382781147957, + "learning_rate": 0.0003896927532837069, + "loss": 3.449807643890381, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.03817906603217125, + "learning_rate": 0.0003894377415479967, + "loss": 3.4556589126586914, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.03523671627044678, + "learning_rate": 0.00038918266017095936, + "loss": 3.477426767349243, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.03284068778157234, + "learning_rate": 0.0003889275093581096, + "loss": 3.478672981262207, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.02992352657020092, + "learning_rate": 0.00038867228931501737, + "loss": 3.4810047149658203, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.02858242578804493, + "learning_rate": 0.0003884170002473091, + "loss": 3.467811107635498, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.027912171557545662, + "learning_rate": 0.0003881616423606666, + "loss": 3.4326171875, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.028351375833153725, + "learning_rate": 0.00038790621586082697, + "loss": 3.4533729553222656, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.027947887778282166, + "learning_rate": 0.0003876507209535829, + "loss": 3.420295476913452, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.028510035946965218, + "learning_rate": 0.00038739515784478186, + "loss": 3.438023328781128, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.027707934379577637, + "learning_rate": 0.00038713952674032654, + "loss": 3.4844181537628174, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.0262764822691679, + "learning_rate": 0.0003868838278461743, + "loss": 3.4363749027252197, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.02713681012392044, + "learning_rate": 0.0003866280613683371, + "loss": 3.4501900672912598, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.0299485195428133, + "learning_rate": 0.00038637222751288126, + "loss": 3.461942672729492, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.027047986164689064, + "learning_rate": 0.0003861163264859277, + "loss": 3.454293727874756, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.02436736971139908, + "learning_rate": 0.0003858603584936511, + "loss": 3.4663405418395996, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.02601189725100994, + "learning_rate": 0.0003856043237422803, + "loss": 3.502030849456787, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.02463468722999096, + "learning_rate": 0.0003853482224380978, + "loss": 3.462491035461426, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.021793190389871597, + "learning_rate": 0.00038509205478744, + "loss": 3.4476377964019775, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.023809107020497322, + "learning_rate": 0.0003848358209966963, + "loss": 3.4701178073883057, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.02557182125747204, + "learning_rate": 0.0003845795212723099, + "loss": 3.4932539463043213, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.023424552753567696, + "learning_rate": 0.00038432315582077664, + "loss": 3.4779105186462402, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.02195148915052414, + "learning_rate": 0.0003840667248486456, + "loss": 3.470506191253662, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.022897250950336456, + "learning_rate": 0.0003838102285625186, + "loss": 3.48982572555542, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.023033805191516876, + "learning_rate": 0.00038355366716905006, + "loss": 3.4845101833343506, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.022774485871195793, + "learning_rate": 0.0003832970408749467, + "loss": 3.5060324668884277, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.02193611115217209, + "learning_rate": 0.00038304034988696816, + "loss": 3.511518955230713, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.02130570448935032, + "learning_rate": 0.00038278359441192516, + "loss": 3.468502998352051, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.020402755588293076, + "learning_rate": 0.00038252677465668136, + "loss": 3.4786908626556396, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.022664198651909828, + "learning_rate": 0.00038226989082815156, + "loss": 3.479928970336914, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.02454056777060032, + "learning_rate": 0.00038201294313330245, + "loss": 3.479947566986084, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.022096732631325722, + "learning_rate": 0.00038175593177915227, + "loss": 3.49478816986084, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.021499406546354294, + "learning_rate": 0.0003814988569727704, + "loss": 3.494933605194092, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.022325703874230385, + "learning_rate": 0.0003812417189212773, + "loss": 3.5041463375091553, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.020674455910921097, + "learning_rate": 0.00038098451783184447, + "loss": 3.472404718399048, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.02131054364144802, + "learning_rate": 0.00038072725391169416, + "loss": 3.497795343399048, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.02126218006014824, + "learning_rate": 0.0003804699273680994, + "loss": 3.483281373977661, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.0217830128967762, + "learning_rate": 0.0003802125384083834, + "loss": 3.4704298973083496, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.02156917378306389, + "learning_rate": 0.0003799550872399197, + "loss": 3.4994399547576904, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.02047055773437023, + "learning_rate": 0.0003796975740701323, + "loss": 3.4316084384918213, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.019766805693507195, + "learning_rate": 0.0003794399991064948, + "loss": 3.452223777770996, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.020028885453939438, + "learning_rate": 0.00037918236255653074, + "loss": 3.485278844833374, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.020312055945396423, + "learning_rate": 0.00037892466462781306, + "loss": 3.4789156913757324, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.020725280046463013, + "learning_rate": 0.00037866690552796446, + "loss": 3.495528221130371, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.02088077738881111, + "learning_rate": 0.00037840908546465696, + "loss": 3.4972569942474365, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.021010037511587143, + "learning_rate": 0.0003781512046456111, + "loss": 3.473562240600586, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.019641095772385597, + "learning_rate": 0.0003778932632785972, + "loss": 3.4782915115356445, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.02063329517841339, + "learning_rate": 0.00037763526157143376, + "loss": 3.4681782722473145, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.0213928259909153, + "learning_rate": 0.00037737719973198825, + "loss": 3.472263813018799, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.02135022170841694, + "learning_rate": 0.0003771190779681762, + "loss": 3.502300262451172, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.02053583413362503, + "learning_rate": 0.0003768608964879619, + "loss": 3.489243745803833, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.021575238555669785, + "learning_rate": 0.00037660265549935724, + "loss": 3.491074562072754, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.022335771471261978, + "learning_rate": 0.00037634435521042257, + "loss": 3.5298843383789062, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.021198464557528496, + "learning_rate": 0.0003760859958292656, + "loss": 3.5156898498535156, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.020986037328839302, + "learning_rate": 0.000375827577564042, + "loss": 3.4574475288391113, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.022639811038970947, + "learning_rate": 0.0003755691006229545, + "loss": 3.4952611923217773, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.019711066037416458, + "learning_rate": 0.0003753105652142534, + "loss": 3.531296730041504, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.02058061584830284, + "learning_rate": 0.000375051971546236, + "loss": 3.5340657234191895, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.019849255681037903, + "learning_rate": 0.0003747933198272465, + "loss": 3.5028295516967773, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01954430714249611, + "learning_rate": 0.00037453461026567604, + "loss": 3.51088809967041, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.019234873354434967, + "learning_rate": 0.00037427584306996196, + "loss": 3.472560167312622, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.021085912361741066, + "learning_rate": 0.0003740170184485888, + "loss": 3.468876838684082, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.020300406962633133, + "learning_rate": 0.0003737581366100864, + "loss": 3.4903886318206787, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02380351349711418, + "learning_rate": 0.0003734991977630315, + "loss": 3.5069704055786133, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.048111915588379, + "eval_runtime": 58.312, + "eval_samples_per_second": 41.878, + "eval_steps_per_second": 1.32, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02379327267408371, + "learning_rate": 0.0003732402021160463, + "loss": 3.396376132965088, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.03204222396016121, + "learning_rate": 0.00037298114987779885, + "loss": 3.4102582931518555, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.034993767738342285, + "learning_rate": 0.000372722041257003, + "loss": 3.385697364807129, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.03387003391981125, + "learning_rate": 0.00037246287646241783, + "loss": 3.4120078086853027, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.03282478079199791, + "learning_rate": 0.0003722036557028478, + "loss": 3.404122829437256, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.03053346462547779, + "learning_rate": 0.0003719443791871422, + "loss": 3.368051767349243, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.029722390696406364, + "learning_rate": 0.0003716850471241958, + "loss": 3.4066247940063477, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.0287226140499115, + "learning_rate": 0.0003714256597229474, + "loss": 3.4212646484375, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.028156550601124763, + "learning_rate": 0.0003711662171923809, + "loss": 3.4192593097686768, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.03021939843893051, + "learning_rate": 0.0003709067197415244, + "loss": 3.3973188400268555, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.03314025700092316, + "learning_rate": 0.00037064716757945036, + "loss": 3.422978401184082, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.03156686946749687, + "learning_rate": 0.0003703875609152753, + "loss": 3.4052343368530273, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.03357478231191635, + "learning_rate": 0.0003701278999581595, + "loss": 3.383730411529541, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.03261081501841545, + "learning_rate": 0.0003698681849173073, + "loss": 3.421250820159912, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.03252755105495453, + "learning_rate": 0.0003696084160019662, + "loss": 3.39884352684021, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.03177274018526077, + "learning_rate": 0.0003693485934214274, + "loss": 3.4175422191619873, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.028845487162470818, + "learning_rate": 0.0003690887173850253, + "loss": 3.3908143043518066, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.026045162230730057, + "learning_rate": 0.0003688287881021374, + "loss": 3.4417710304260254, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.027937524020671844, + "learning_rate": 0.00036856880578218376, + "loss": 3.410548210144043, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.025780271738767624, + "learning_rate": 0.00036830877063462786, + "loss": 3.4069018363952637, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.026093080639839172, + "learning_rate": 0.0003680486828689749, + "loss": 3.4199600219726562, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.025765180587768555, + "learning_rate": 0.00036778854269477315, + "loss": 3.442903995513916, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.022773101925849915, + "learning_rate": 0.0003675283503216127, + "loss": 3.394232749938965, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.025135623291134834, + "learning_rate": 0.000367268105959126, + "loss": 3.4466171264648438, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.02491890639066696, + "learning_rate": 0.00036700780981698705, + "loss": 3.418689727783203, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.02521841786801815, + "learning_rate": 0.0003667474621049119, + "loss": 3.4213061332702637, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.025186199694871902, + "learning_rate": 0.00036648706303265795, + "loss": 3.437631130218506, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.022097431123256683, + "learning_rate": 0.000366226612810024, + "loss": 3.4014902114868164, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.024009618908166885, + "learning_rate": 0.00036596611164685025, + "loss": 3.4292187690734863, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.022758696228265762, + "learning_rate": 0.0003657055597530175, + "loss": 3.4105381965637207, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.02447321265935898, + "learning_rate": 0.00036544495733844797, + "loss": 3.4196958541870117, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.025843804702162743, + "learning_rate": 0.000365184304613104, + "loss": 3.4343783855438232, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.022433951497077942, + "learning_rate": 0.0003649236017869892, + "loss": 3.440580368041992, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.02250702679157257, + "learning_rate": 0.00036466284907014685, + "loss": 3.4255261421203613, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.022953296080231667, + "learning_rate": 0.0003644020466726608, + "loss": 3.442258358001709, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.021977661177515984, + "learning_rate": 0.00036414119480465473, + "loss": 3.4058687686920166, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.021751178428530693, + "learning_rate": 0.0003638802936762925, + "loss": 3.4211649894714355, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.022764205932617188, + "learning_rate": 0.0003636193434977771, + "loss": 3.424694061279297, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.024827787652611732, + "learning_rate": 0.0003633583444793516, + "loss": 3.458076238632202, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.021888602524995804, + "learning_rate": 0.00036309729683129814, + "loss": 3.4509153366088867, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.021961702033877373, + "learning_rate": 0.00036283620076393814, + "loss": 3.4147253036499023, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.02235778421163559, + "learning_rate": 0.000362575056487632, + "loss": 3.4392640590667725, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.021714260801672935, + "learning_rate": 0.00036231386421277877, + "loss": 3.444026470184326, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.020996905863285065, + "learning_rate": 0.0003620526241498167, + "loss": 3.4168434143066406, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.020528919994831085, + "learning_rate": 0.0003617913365092218, + "loss": 3.4831130504608154, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.021056141704320908, + "learning_rate": 0.0003615300015015091, + "loss": 3.430905342102051, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.021655110642313957, + "learning_rate": 0.0003612686193372312, + "loss": 3.4588241577148438, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.02344612590968609, + "learning_rate": 0.00036100719022697924, + "loss": 3.4202680587768555, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.02264723740518093, + "learning_rate": 0.0003607457143813819, + "loss": 3.4669015407562256, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.02377854660153389, + "learning_rate": 0.00036048419201110553, + "loss": 3.5059216022491455, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.02299923449754715, + "learning_rate": 0.00036022262332685383, + "loss": 3.4796524047851562, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.021110547706484795, + "learning_rate": 0.0003599610085393681, + "loss": 3.4342408180236816, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.022060122340917587, + "learning_rate": 0.0003596993478594267, + "loss": 3.426923990249634, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.021077830344438553, + "learning_rate": 0.0003594376414978447, + "loss": 3.4636270999908447, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.02209082990884781, + "learning_rate": 0.0003591758896654745, + "loss": 3.441979169845581, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.022408071905374527, + "learning_rate": 0.0003589140925732045, + "loss": 3.490835666656494, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.022462666034698486, + "learning_rate": 0.0003586522504319602, + "loss": 3.452770233154297, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.021295800805091858, + "learning_rate": 0.0003583903634527029, + "loss": 3.4377408027648926, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.020457757636904716, + "learning_rate": 0.0003581284318464302, + "loss": 3.4191250801086426, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.021163633093237877, + "learning_rate": 0.00035786645582417564, + "loss": 3.4797091484069824, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.021495461463928223, + "learning_rate": 0.00035760443559700863, + "loss": 3.4763145446777344, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.021050071343779564, + "learning_rate": 0.00035734237137603417, + "loss": 3.4687209129333496, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.02150929905474186, + "learning_rate": 0.0003570802633723927, + "loss": 3.4819254875183105, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.0212967898696661, + "learning_rate": 0.0003568181117972597, + "loss": 3.490407705307007, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.02123691327869892, + "learning_rate": 0.00035655591686184626, + "loss": 3.4418535232543945, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.02193271741271019, + "learning_rate": 0.0003562936787773979, + "loss": 3.446892738342285, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.02065582573413849, + "learning_rate": 0.000356031397755195, + "loss": 3.4308691024780273, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.019487040117383003, + "learning_rate": 0.00035576907400655303, + "loss": 3.487239122390747, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.021190891042351723, + "learning_rate": 0.00035550670774282106, + "loss": 3.4635558128356934, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.0209872554987669, + "learning_rate": 0.0003552442991753831, + "loss": 3.443793773651123, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.02035645768046379, + "learning_rate": 0.00035498184851565697, + "loss": 3.473198652267456, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.024002201855182648, + "learning_rate": 0.0003547193559750944, + "loss": 3.4244751930236816, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.057085990905762, + "eval_runtime": 62.4735, + "eval_samples_per_second": 39.089, + "eval_steps_per_second": 1.233, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.02450242079794407, + "learning_rate": 0.00035445682176518087, + "loss": 3.361237049102783, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.030813416466116905, + "learning_rate": 0.00035419424609743546, + "loss": 3.365708351135254, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.03279578685760498, + "learning_rate": 0.00035393162918341065, + "loss": 3.325385570526123, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.03293970227241516, + "learning_rate": 0.00035366897123469213, + "loss": 3.356811285018921, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.033190157264471054, + "learning_rate": 0.0003534062724628986, + "loss": 3.406217098236084, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.033370476216077805, + "learning_rate": 0.00035314353307968195, + "loss": 3.3579213619232178, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.03351330757141113, + "learning_rate": 0.00035288075329672634, + "loss": 3.3755106925964355, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.034596871584653854, + "learning_rate": 0.0003526179333257487, + "loss": 3.3576536178588867, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.03177076578140259, + "learning_rate": 0.0003523550733784984, + "loss": 3.3123257160186768, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.0289607010781765, + "learning_rate": 0.00035209217366675706, + "loss": 3.370959997177124, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.03265267238020897, + "learning_rate": 0.000351829234402338, + "loss": 3.332608938217163, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.03397805988788605, + "learning_rate": 0.0003515662557970867, + "loss": 3.364213228225708, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.03100395016372204, + "learning_rate": 0.0003513032380628804, + "loss": 3.388005256652832, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.03224753588438034, + "learning_rate": 0.0003510401814116277, + "loss": 3.3267595767974854, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.03034576214849949, + "learning_rate": 0.0003507770860552684, + "loss": 3.3954129219055176, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.026316817849874496, + "learning_rate": 0.00035051395220577397, + "loss": 3.3543832302093506, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.027714375406503677, + "learning_rate": 0.0003502507800751464, + "loss": 3.4190099239349365, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.028731998056173325, + "learning_rate": 0.0003499875698754187, + "loss": 3.342031955718994, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02845810353755951, + "learning_rate": 0.00034972432181865467, + "loss": 3.356130599975586, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.02842753194272518, + "learning_rate": 0.00034946103611694854, + "loss": 3.3709869384765625, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.02942183054983616, + "learning_rate": 0.00034919771298242477, + "loss": 3.373588800430298, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.02958574891090393, + "learning_rate": 0.00034893435262723814, + "loss": 3.377514600753784, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.029524607583880424, + "learning_rate": 0.00034867095526357325, + "loss": 3.398663282394409, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.028092598542571068, + "learning_rate": 0.0003484075211036446, + "loss": 3.3742518424987793, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.026652829721570015, + "learning_rate": 0.0003481440503596964, + "loss": 3.387667179107666, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.024675756692886353, + "learning_rate": 0.0003478805432440021, + "loss": 3.3737070560455322, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.02491692826151848, + "learning_rate": 0.0003476169999688648, + "loss": 3.359217643737793, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.024601681157946587, + "learning_rate": 0.00034735342074661654, + "loss": 3.3902688026428223, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.02313913404941559, + "learning_rate": 0.0003470898057896183, + "loss": 3.3700339794158936, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.02418898604810238, + "learning_rate": 0.0003468261553102599, + "loss": 3.4265692234039307, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.023382030427455902, + "learning_rate": 0.00034656246952095984, + "loss": 3.378389835357666, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.02340470254421234, + "learning_rate": 0.0003462987486341648, + "loss": 3.392937421798706, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.022166339680552483, + "learning_rate": 0.00034603499286235006, + "loss": 3.39414119720459, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.022175129503011703, + "learning_rate": 0.0003457712024180188, + "loss": 3.407447099685669, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.02368207648396492, + "learning_rate": 0.0003455073775137025, + "loss": 3.415076971054077, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.02263982594013214, + "learning_rate": 0.0003452435183619598, + "loss": 3.383986473083496, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.02147712931036949, + "learning_rate": 0.0003449796251753773, + "loss": 3.403963565826416, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.022987166419625282, + "learning_rate": 0.00034471569816656915, + "loss": 3.3883113861083984, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.02240179106593132, + "learning_rate": 0.00034445173754817646, + "loss": 3.4349911212921143, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.02206592820584774, + "learning_rate": 0.00034418774353286747, + "loss": 3.426473617553711, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.022650036960840225, + "learning_rate": 0.0003439237163333375, + "loss": 3.4251773357391357, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.023399339988827705, + "learning_rate": 0.0003436596561623084, + "loss": 3.371610641479492, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.02192695252597332, + "learning_rate": 0.0003433955632325288, + "loss": 3.387253761291504, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.02241935208439827, + "learning_rate": 0.00034313143775677353, + "loss": 3.434523105621338, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.023202672600746155, + "learning_rate": 0.00034286727994784367, + "loss": 3.369771957397461, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.021333014592528343, + "learning_rate": 0.0003426030900185665, + "loss": 3.439004421234131, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.023988714441657066, + "learning_rate": 0.0003423388681817949, + "loss": 3.396838665008545, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.023885777220129967, + "learning_rate": 0.00034207461465040793, + "loss": 3.4114322662353516, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.02263713628053665, + "learning_rate": 0.0003418103296373096, + "loss": 3.4139351844787598, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.022404303774237633, + "learning_rate": 0.00034154601335542964, + "loss": 3.4200823307037354, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.021705076098442078, + "learning_rate": 0.00034128166601772304, + "loss": 3.4092609882354736, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.02273249812424183, + "learning_rate": 0.0003410172878371695, + "loss": 3.3931846618652344, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.022552402690052986, + "learning_rate": 0.00034075287902677394, + "loss": 3.396566867828369, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.023219924420118332, + "learning_rate": 0.0003404884397995655, + "loss": 3.369025707244873, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.022560900077223778, + "learning_rate": 0.00034022397036859837, + "loss": 3.4248225688934326, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.023370858281850815, + "learning_rate": 0.0003399594709469506, + "loss": 3.431628942489624, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.02469109185039997, + "learning_rate": 0.00033969494174772465, + "loss": 3.406843662261963, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.02417670376598835, + "learning_rate": 0.0003394303829840469, + "loss": 3.445046901702881, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.0237257182598114, + "learning_rate": 0.00033916579486906764, + "loss": 3.4490597248077393, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.02295520156621933, + "learning_rate": 0.0003389011776159607, + "loss": 3.4009876251220703, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.023158475756645203, + "learning_rate": 0.0003386365314379233, + "loss": 3.4383902549743652, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.023618478327989578, + "learning_rate": 0.0003383718565481763, + "loss": 3.421168804168701, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.022055329754948616, + "learning_rate": 0.0003381071531599633, + "loss": 3.382946729660034, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.021441346034407616, + "learning_rate": 0.00033784242148655115, + "loss": 3.4464712142944336, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.023031961172819138, + "learning_rate": 0.00033757766174122934, + "loss": 3.42844295501709, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.022832952439785004, + "learning_rate": 0.00033731287413731005, + "loss": 3.398231029510498, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.022094765678048134, + "learning_rate": 0.0003370480588881278, + "loss": 3.4270074367523193, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.0218376275151968, + "learning_rate": 0.0003367832162070395, + "loss": 3.4681742191314697, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.02364739589393139, + "learning_rate": 0.000336518346307424, + "loss": 3.431920051574707, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.023402225226163864, + "learning_rate": 0.0003362534494026824, + "loss": 3.4189023971557617, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.022775836288928986, + "learning_rate": 0.0003359885257062372, + "loss": 3.409396171569824, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.02445383556187153, + "learning_rate": 0.0003357235754315328, + "loss": 3.41017746925354, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.066352367401123, + "eval_runtime": 58.1658, + "eval_samples_per_second": 41.983, + "eval_steps_per_second": 1.324, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.02710847556591034, + "learning_rate": 0.0003354585987920345, + "loss": 3.3155131340026855, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.03428799286484718, + "learning_rate": 0.0003351935960012296, + "loss": 3.327408790588379, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.03434379771351814, + "learning_rate": 0.0003349285672726259, + "loss": 3.3282251358032227, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.03405233472585678, + "learning_rate": 0.0003346635128197522, + "loss": 3.2932448387145996, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.031194346025586128, + "learning_rate": 0.0003343984328561581, + "loss": 3.336615800857544, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.031816478818655014, + "learning_rate": 0.00033413332759541376, + "loss": 3.3375275135040283, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.03413299471139908, + "learning_rate": 0.0003338681972511098, + "loss": 3.3710694313049316, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.034200165420770645, + "learning_rate": 0.0003336030420368568, + "loss": 3.3248214721679688, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.030694160610437393, + "learning_rate": 0.0003333378621662857, + "loss": 3.3297901153564453, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.031172804534435272, + "learning_rate": 0.00033307265785304684, + "loss": 3.287217140197754, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.031914614140987396, + "learning_rate": 0.00033280742931081076, + "loss": 3.3218531608581543, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.0332055389881134, + "learning_rate": 0.00033254217675326737, + "loss": 3.3454785346984863, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.034956980496644974, + "learning_rate": 0.0003322769003941257, + "loss": 3.3193702697753906, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.03618055582046509, + "learning_rate": 0.00033201160044711423, + "loss": 3.3661017417907715, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.030322827398777008, + "learning_rate": 0.0003317462771259802, + "loss": 3.353989601135254, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.029768960550427437, + "learning_rate": 0.00033148093064449006, + "loss": 3.3254497051239014, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.028036782518029213, + "learning_rate": 0.0003312155612164284, + "loss": 3.3276920318603516, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.028774483129382133, + "learning_rate": 0.00033095016905559883, + "loss": 3.3458826541900635, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.030403515323996544, + "learning_rate": 0.0003306847543758227, + "loss": 3.3332793712615967, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.028468478471040726, + "learning_rate": 0.0003304193173909401, + "loss": 3.3167755603790283, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.029126476496458054, + "learning_rate": 0.00033015385831480873, + "loss": 3.299426794052124, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.030746638774871826, + "learning_rate": 0.0003298883773613043, + "loss": 3.3581809997558594, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.030241351574659348, + "learning_rate": 0.0003296228747443197, + "loss": 3.3415942192077637, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.029754415154457092, + "learning_rate": 0.0003293573506777659, + "loss": 3.3525662422180176, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.026896746829152107, + "learning_rate": 0.00032909180537557076, + "loss": 3.321808099746704, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.026361066848039627, + "learning_rate": 0.00032882623905167917, + "loss": 3.329258441925049, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.027747981250286102, + "learning_rate": 0.00032856065192005335, + "loss": 3.3774523735046387, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.026270100846886635, + "learning_rate": 0.00032829504419467194, + "loss": 3.3419976234436035, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.02731695957481861, + "learning_rate": 0.0003280294160895303, + "loss": 3.3068814277648926, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.026583299040794373, + "learning_rate": 0.0003277637678186402, + "loss": 3.303804874420166, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.026615949347615242, + "learning_rate": 0.00032749809959602973, + "loss": 3.3812599182128906, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.027221621945500374, + "learning_rate": 0.0003272324116357428, + "loss": 3.327211618423462, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.025025667622685432, + "learning_rate": 0.0003269667041518395, + "loss": 3.339163303375244, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.02593836560845375, + "learning_rate": 0.0003267009773583956, + "loss": 3.3517372608184814, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.023697076365351677, + "learning_rate": 0.00032643523146950234, + "loss": 3.379819869995117, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.02425902709364891, + "learning_rate": 0.0003261694666992664, + "loss": 3.390653610229492, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.025133250281214714, + "learning_rate": 0.0003259036832618096, + "loss": 3.364495277404785, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.025542905554175377, + "learning_rate": 0.0003256378813712688, + "loss": 3.3584814071655273, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.026041125878691673, + "learning_rate": 0.000325372061241796, + "loss": 3.36777400970459, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.02391037344932556, + "learning_rate": 0.00032510622308755746, + "loss": 3.3809385299682617, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.02379394695162773, + "learning_rate": 0.00032484036712273424, + "loss": 3.354257583618164, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.0234722550958395, + "learning_rate": 0.0003245744935615219, + "loss": 3.3736205101013184, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.024123650044202805, + "learning_rate": 0.0003243086026181296, + "loss": 3.3589959144592285, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.024013696238398552, + "learning_rate": 0.00032404269450678116, + "loss": 3.353874444961548, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.022617843002080917, + "learning_rate": 0.00032377676944171375, + "loss": 3.364056348800659, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.022730547934770584, + "learning_rate": 0.00032351082763717857, + "loss": 3.378067970275879, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.022257911041378975, + "learning_rate": 0.0003232448693074399, + "loss": 3.379342555999756, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.023247094824910164, + "learning_rate": 0.00032297889466677575, + "loss": 3.3537116050720215, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.022764021530747414, + "learning_rate": 0.000322712903929477, + "loss": 3.3969063758850098, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.02260403335094452, + "learning_rate": 0.0003224468973098477, + "loss": 3.3727574348449707, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.021772351115942, + "learning_rate": 0.0003221808750222044, + "loss": 3.335434675216675, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.02238459885120392, + "learning_rate": 0.0003219148372808766, + "loss": 3.3440277576446533, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.0230540968477726, + "learning_rate": 0.00032164878430020606, + "loss": 3.3997583389282227, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.0220821350812912, + "learning_rate": 0.00032138271629454684, + "loss": 3.377815008163452, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.021497417241334915, + "learning_rate": 0.00032111663347826505, + "loss": 3.3696370124816895, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.02163885347545147, + "learning_rate": 0.00032085053606573896, + "loss": 3.3489794731140137, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.022394269704818726, + "learning_rate": 0.0003205844242713584, + "loss": 3.3792812824249268, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.022390708327293396, + "learning_rate": 0.0003203182983095248, + "loss": 3.3733272552490234, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.023712158203125, + "learning_rate": 0.0003200521583946511, + "loss": 3.3716251850128174, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.02398860827088356, + "learning_rate": 0.0003197860047411613, + "loss": 3.381779670715332, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.02309161052107811, + "learning_rate": 0.00031951983756349066, + "loss": 3.3923277854919434, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.021400388330221176, + "learning_rate": 0.00031925365707608536, + "loss": 3.407397747039795, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.022534485906362534, + "learning_rate": 0.00031898746349340204, + "loss": 3.3740391731262207, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.023533202707767487, + "learning_rate": 0.0003187212570299082, + "loss": 3.3877944946289062, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.023398835211992264, + "learning_rate": 0.0003184550379000815, + "loss": 3.3936471939086914, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.020927201956510544, + "learning_rate": 0.00031818880631841005, + "loss": 3.339545249938965, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.021453995257616043, + "learning_rate": 0.0003179225624993915, + "loss": 3.3647522926330566, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.022847114130854607, + "learning_rate": 0.0003176563066575341, + "loss": 3.405144691467285, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.022766996175050735, + "learning_rate": 0.0003173900390073549, + "loss": 3.392810344696045, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.022103948518633842, + "learning_rate": 0.0003171237597633813, + "loss": 3.3973004817962646, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.022282181307673454, + "learning_rate": 0.0003168574691401496, + "loss": 3.409998893737793, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.025435522198677063, + "learning_rate": 0.0003165911673522053, + "loss": 3.423318386077881, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.078423500061035, + "eval_runtime": 57.3189, + "eval_samples_per_second": 42.604, + "eval_steps_per_second": 1.343, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.02624760940670967, + "learning_rate": 0.00031632485461410294, + "loss": 3.261288642883301, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.03552987053990364, + "learning_rate": 0.00031605853114040583, + "loss": 3.2682106494903564, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.040380869060754776, + "learning_rate": 0.00031579219714568613, + "loss": 3.2813782691955566, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.03861293941736221, + "learning_rate": 0.0003155258528445242, + "loss": 3.258193016052246, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.04091731086373329, + "learning_rate": 0.0003152594984515089, + "loss": 3.3052220344543457, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.043269723653793335, + "learning_rate": 0.00031499313418123704, + "loss": 3.2820217609405518, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.04580320790410042, + "learning_rate": 0.0003147267602483138, + "loss": 3.312005043029785, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.047336459159851074, + "learning_rate": 0.00031446037686735153, + "loss": 3.3217713832855225, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.04587360844016075, + "learning_rate": 0.0003141939842529706, + "loss": 3.2869949340820312, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.04002532362937927, + "learning_rate": 0.0003139275826197987, + "loss": 3.2879223823547363, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.0368693508207798, + "learning_rate": 0.00031366117218247094, + "loss": 3.2983577251434326, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.037317775189876556, + "learning_rate": 0.00031339475315562916, + "loss": 3.3062524795532227, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.03655700013041496, + "learning_rate": 0.00031312832575392267, + "loss": 3.31072735786438, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.033466216176748276, + "learning_rate": 0.0003128618901920071, + "loss": 3.2954514026641846, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.030785970389842987, + "learning_rate": 0.0003125954466845446, + "loss": 3.27952241897583, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.029669342562556267, + "learning_rate": 0.00031232899544620404, + "loss": 3.2613229751586914, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.033505845814943314, + "learning_rate": 0.00031206253669166045, + "loss": 3.2987258434295654, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.032395437359809875, + "learning_rate": 0.0003117960706355947, + "loss": 3.296708583831787, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.031322430819272995, + "learning_rate": 0.0003115295974926936, + "loss": 3.3160252571105957, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.029345186427235603, + "learning_rate": 0.00031126311747765005, + "loss": 3.307143211364746, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.026715675368905067, + "learning_rate": 0.00031099663080516186, + "loss": 3.286012649536133, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.028592750430107117, + "learning_rate": 0.0003107301376899327, + "loss": 3.2528645992279053, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.028861764818429947, + "learning_rate": 0.00031046363834667123, + "loss": 3.2795748710632324, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.02609528973698616, + "learning_rate": 0.00031019713299009123, + "loss": 3.326317071914673, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.027718454599380493, + "learning_rate": 0.000309930621834911, + "loss": 3.268110513687134, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.026382867246866226, + "learning_rate": 0.0003096641050958541, + "loss": 3.294499397277832, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.027083583176136017, + "learning_rate": 0.00030939758298764795, + "loss": 3.3202130794525146, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.02516467310488224, + "learning_rate": 0.00030913105572502483, + "loss": 3.3485021591186523, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.024148069322109222, + "learning_rate": 0.00030886452352272095, + "loss": 3.3099541664123535, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.024552857503294945, + "learning_rate": 0.00030859798659547624, + "loss": 3.2885942459106445, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.02551114559173584, + "learning_rate": 0.0003083314451580349, + "loss": 3.3232040405273438, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.025423647835850716, + "learning_rate": 0.0003080648994251445, + "loss": 3.3097832202911377, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.023211417719721794, + "learning_rate": 0.0003077983496115561, + "loss": 3.3075966835021973, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.024623898789286613, + "learning_rate": 0.00030753179593202406, + "loss": 3.2917377948760986, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.02677663415670395, + "learning_rate": 0.0003072652386013059, + "loss": 3.356265068054199, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.02673642337322235, + "learning_rate": 0.00030699867783416207, + "loss": 3.314903736114502, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.024466225877404213, + "learning_rate": 0.0003067321138453557, + "loss": 3.30912446975708, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.025370748713612556, + "learning_rate": 0.0003064655468496527, + "loss": 3.363640069961548, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.026536114513874054, + "learning_rate": 0.0003061989770618211, + "loss": 3.304851531982422, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.026536893099546432, + "learning_rate": 0.00030593240469663154, + "loss": 3.342902898788452, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.02645307406783104, + "learning_rate": 0.0003056658299688563, + "loss": 3.3164432048797607, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.02523060515522957, + "learning_rate": 0.00030539925309327017, + "loss": 3.34376859664917, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.02590431459248066, + "learning_rate": 0.0003051326742846491, + "loss": 3.3479061126708984, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.0230655986815691, + "learning_rate": 0.000304866093757771, + "loss": 3.2890982627868652, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.02403893694281578, + "learning_rate": 0.00030459951172741474, + "loss": 3.345456600189209, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.02238907851278782, + "learning_rate": 0.0003043329284083609, + "loss": 3.3471767902374268, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.021377569064497948, + "learning_rate": 0.00030406634401539053, + "loss": 3.319232225418091, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.02272353693842888, + "learning_rate": 0.00030379975876328614, + "loss": 3.316892623901367, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.02413824200630188, + "learning_rate": 0.00030353317286683053, + "loss": 3.3450968265533447, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.02334682270884514, + "learning_rate": 0.0003032665865408073, + "loss": 3.3640642166137695, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.021966520696878433, + "learning_rate": 0.000303, + "loss": 3.358555316925049, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.023984894156455994, + "learning_rate": 0.0003027334134591926, + "loss": 3.358398914337158, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.023520268499851227, + "learning_rate": 0.00030246682713316945, + "loss": 3.3224587440490723, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.0231285709887743, + "learning_rate": 0.0003022002412367138, + "loss": 3.3671562671661377, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.022818170487880707, + "learning_rate": 0.00030193365598460946, + "loss": 3.361027717590332, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.02280694991350174, + "learning_rate": 0.0003016670715916391, + "loss": 3.2923383712768555, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.021362345665693283, + "learning_rate": 0.00030140048827258524, + "loss": 3.3346734046936035, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.02372218668460846, + "learning_rate": 0.0003011339062422289, + "loss": 3.354748249053955, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.022949764505028725, + "learning_rate": 0.00030086732571535076, + "loss": 3.3857789039611816, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.021954255178570747, + "learning_rate": 0.0003006007469067297, + "loss": 3.343330144882202, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.022438595071434975, + "learning_rate": 0.0003003341700311436, + "loss": 3.328618049621582, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.021969925612211227, + "learning_rate": 0.0003000675953033685, + "loss": 3.355971574783325, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.022706981748342514, + "learning_rate": 0.0002998010229381789, + "loss": 3.3529577255249023, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.024083172902464867, + "learning_rate": 0.0002995344531503474, + "loss": 3.348170757293701, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.023259354755282402, + "learning_rate": 0.0002992678861546442, + "loss": 3.3474369049072266, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.02218157798051834, + "learning_rate": 0.00029900132216583786, + "loss": 3.369053840637207, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.02305331453680992, + "learning_rate": 0.0002987347613986941, + "loss": 3.3622641563415527, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.022796178236603737, + "learning_rate": 0.0002984682040679759, + "loss": 3.351510524749756, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.021299205720424652, + "learning_rate": 0.00029820165038844395, + "loss": 3.339406728744507, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.022786924615502357, + "learning_rate": 0.00029793510057485556, + "loss": 3.3182196617126465, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.021729128435254097, + "learning_rate": 0.00029766855484196513, + "loss": 3.383904218673706, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.024604283273220062, + "learning_rate": 0.0002974020134045238, + "loss": 3.2912864685058594, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.0894880294799805, + "eval_runtime": 58.3967, + "eval_samples_per_second": 41.817, + "eval_steps_per_second": 1.319, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.02673362009227276, + "learning_rate": 0.00029713547647727904, + "loss": 3.270986795425415, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.04036599025130272, + "learning_rate": 0.00029686894427497515, + "loss": 3.2358999252319336, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.047250088304281235, + "learning_rate": 0.0002966024170123519, + "loss": 3.2706189155578613, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.04642819985747337, + "learning_rate": 0.0002963358949041459, + "loss": 3.236340284347534, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.039405327290296555, + "learning_rate": 0.00029606937816508897, + "loss": 3.203136682510376, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.038748081773519516, + "learning_rate": 0.00029580286700990887, + "loss": 3.2636070251464844, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.039233580231666565, + "learning_rate": 0.00029553636165332876, + "loss": 3.2751896381378174, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.03429320082068443, + "learning_rate": 0.0002952698623100673, + "loss": 3.2765679359436035, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.03267038241028786, + "learning_rate": 0.0002950033691948382, + "loss": 3.2387681007385254, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.03349406272172928, + "learning_rate": 0.00029473688252234994, + "loss": 3.282162666320801, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.032002877444028854, + "learning_rate": 0.0002944704025073064, + "loss": 3.246623992919922, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.031509071588516235, + "learning_rate": 0.00029420392936440525, + "loss": 3.2374932765960693, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.02958357334136963, + "learning_rate": 0.00029393746330833953, + "loss": 3.24362850189209, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.03248635679483414, + "learning_rate": 0.0002936710045537959, + "loss": 3.237210273742676, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.03230655938386917, + "learning_rate": 0.0002934045533154554, + "loss": 3.251129150390625, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.029688173905014992, + "learning_rate": 0.00029313810980799296, + "loss": 3.2141575813293457, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.027773940935730934, + "learning_rate": 0.00029287167424607726, + "loss": 3.264249086380005, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.02687031403183937, + "learning_rate": 0.00029260524684437077, + "loss": 3.2513632774353027, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.02503623627126217, + "learning_rate": 0.00029233882781752905, + "loss": 3.217961072921753, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.027071228250861168, + "learning_rate": 0.00029207241738020136, + "loss": 3.2784645557403564, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.02600773423910141, + "learning_rate": 0.0002918060157470294, + "loss": 3.2641382217407227, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.026998162269592285, + "learning_rate": 0.00029153962313264856, + "loss": 3.2646782398223877, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.025579402223229408, + "learning_rate": 0.0002912732397516862, + "loss": 3.2700581550598145, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.02645333670079708, + "learning_rate": 0.00029100686581876283, + "loss": 3.2545266151428223, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.02717350795865059, + "learning_rate": 0.0002907405015484911, + "loss": 3.2795705795288086, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.02675914391875267, + "learning_rate": 0.0002904741471554758, + "loss": 3.2936904430389404, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.02751768007874489, + "learning_rate": 0.0002902078028543139, + "loss": 3.2329440116882324, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.025446882471442223, + "learning_rate": 0.00028994146885959416, + "loss": 3.274294376373291, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.02365768514573574, + "learning_rate": 0.00028967514538589715, + "loss": 3.2585229873657227, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.0266884732991457, + "learning_rate": 0.0002894088326477947, + "loss": 3.2872962951660156, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.02572469227015972, + "learning_rate": 0.0002891425308598503, + "loss": 3.253828287124634, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.02540273219347, + "learning_rate": 0.00028887624023661866, + "loss": 3.303926944732666, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.024694625288248062, + "learning_rate": 0.00028860996099264505, + "loss": 3.264145851135254, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.027233170345425606, + "learning_rate": 0.000288343693342466, + "loss": 3.286924362182617, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02635285072028637, + "learning_rate": 0.00028807743750060847, + "loss": 3.2930331230163574, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.025157805532217026, + "learning_rate": 0.00028781119368159004, + "loss": 3.2906906604766846, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.024729596450924873, + "learning_rate": 0.00028754496209991843, + "loss": 3.277128219604492, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.023809000849723816, + "learning_rate": 0.0002872787429700917, + "loss": 3.2957217693328857, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02474421076476574, + "learning_rate": 0.0002870125365065979, + "loss": 3.27791690826416, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.02439294196665287, + "learning_rate": 0.00028674634292391463, + "loss": 3.2912981510162354, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.023601334542036057, + "learning_rate": 0.0002864801624365093, + "loss": 3.317446231842041, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.0253264419734478, + "learning_rate": 0.0002862139952588387, + "loss": 3.2742302417755127, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.02389134280383587, + "learning_rate": 0.000285947841605349, + "loss": 3.281032085418701, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.02451813407242298, + "learning_rate": 0.00028568170169047523, + "loss": 3.3339295387268066, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.024324845522642136, + "learning_rate": 0.00028541557572864154, + "loss": 3.290402889251709, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.024939008057117462, + "learning_rate": 0.000285149463934261, + "loss": 3.2933621406555176, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.024450505152344704, + "learning_rate": 0.0002848833665217349, + "loss": 3.3055014610290527, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.02262270078063011, + "learning_rate": 0.0002846172837054532, + "loss": 3.2964205741882324, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.02311946265399456, + "learning_rate": 0.0002843512156997939, + "loss": 3.3102402687072754, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.024362705647945404, + "learning_rate": 0.0002840851627191234, + "loss": 3.3020925521850586, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.024785032495856285, + "learning_rate": 0.0002838191249777955, + "loss": 3.287290096282959, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.02358846180140972, + "learning_rate": 0.00028355310269015225, + "loss": 3.334261894226074, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.024330584332346916, + "learning_rate": 0.00028328709607052297, + "loss": 3.2799081802368164, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.023546984419226646, + "learning_rate": 0.0002830211053332242, + "loss": 3.3228847980499268, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.023060832172632217, + "learning_rate": 0.0002827551306925601, + "loss": 3.325793981552124, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.02388455718755722, + "learning_rate": 0.00028248917236282147, + "loss": 3.2892062664031982, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.023740701377391815, + "learning_rate": 0.0002822232305582863, + "loss": 3.2969369888305664, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.02458217367529869, + "learning_rate": 0.0002819573054932188, + "loss": 3.315610408782959, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.023173579946160316, + "learning_rate": 0.0002816913973818703, + "loss": 3.2561936378479004, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.024137936532497406, + "learning_rate": 0.0002814255064384781, + "loss": 3.2839202880859375, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.022436272352933884, + "learning_rate": 0.00028115963287726563, + "loss": 3.3057351112365723, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.023997878655791283, + "learning_rate": 0.0002808937769124425, + "loss": 3.3583154678344727, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.024119896814227104, + "learning_rate": 0.000280627938758204, + "loss": 3.322892427444458, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.022283365949988365, + "learning_rate": 0.00028036211862873117, + "loss": 3.3145735263824463, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.023433363065123558, + "learning_rate": 0.0002800963167381904, + "loss": 3.3564529418945312, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.02224746160209179, + "learning_rate": 0.00027983053330073355, + "loss": 3.2985916137695312, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.023786570876836777, + "learning_rate": 0.00027956476853049765, + "loss": 3.302539587020874, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.025200961157679558, + "learning_rate": 0.00027929902264160435, + "loss": 3.309976577758789, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.023849809542298317, + "learning_rate": 0.00027903329584816046, + "loss": 3.343660354614258, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.023311031982302666, + "learning_rate": 0.0002787675883642572, + "loss": 3.2978909015655518, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.023174891248345375, + "learning_rate": 0.00027850190040397036, + "loss": 3.326188802719116, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.02681034244596958, + "learning_rate": 0.00027823623218135976, + "loss": 3.3422017097473145, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.109009265899658, + "eval_runtime": 58.0506, + "eval_samples_per_second": 42.067, + "eval_steps_per_second": 1.326, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.028057295829057693, + "learning_rate": 0.0002779705839104697, + "loss": 3.1825037002563477, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.034786805510520935, + "learning_rate": 0.0002777049558053281, + "loss": 3.200389862060547, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.03751441463828087, + "learning_rate": 0.0002774393480799466, + "loss": 3.2047512531280518, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.035590436309576035, + "learning_rate": 0.00027717376094832076, + "loss": 3.1735124588012695, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.03426426649093628, + "learning_rate": 0.0002769081946244293, + "loss": 3.199483871459961, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.037030890583992004, + "learning_rate": 0.00027664264932223407, + "loss": 3.213322639465332, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.032468877732753754, + "learning_rate": 0.00027637712525568024, + "loss": 3.179354190826416, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.027202758938074112, + "learning_rate": 0.0002761116226386958, + "loss": 3.216036319732666, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.031184563413262367, + "learning_rate": 0.00027584614168519125, + "loss": 3.208205223083496, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.029414812102913857, + "learning_rate": 0.00027558068260905976, + "loss": 3.180370569229126, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.02679901383817196, + "learning_rate": 0.00027531524562417734, + "loss": 3.210325002670288, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.02863631211221218, + "learning_rate": 0.00027504983094440126, + "loss": 3.1601076126098633, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.029135366901755333, + "learning_rate": 0.0002747844387835716, + "loss": 3.2262887954711914, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.030535712838172913, + "learning_rate": 0.0002745190693555099, + "loss": 3.23160719871521, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.030563877895474434, + "learning_rate": 0.0002742537228740197, + "loss": 3.255211353302002, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.02906505949795246, + "learning_rate": 0.00027398839955288575, + "loss": 3.2006797790527344, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.030871668830513954, + "learning_rate": 0.00027372309960587424, + "loss": 3.2442867755889893, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.02913421392440796, + "learning_rate": 0.0002734578232467327, + "loss": 3.218442440032959, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.030087953433394432, + "learning_rate": 0.0002731925706891891, + "loss": 3.2164535522460938, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.034147944301366806, + "learning_rate": 0.0002729273421469532, + "loss": 3.218660831451416, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.03593966364860535, + "learning_rate": 0.00027266213783371434, + "loss": 3.2240614891052246, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.032090652734041214, + "learning_rate": 0.00027239695796314327, + "loss": 3.2309441566467285, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.02776424027979374, + "learning_rate": 0.0002721318027488902, + "loss": 3.2702016830444336, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.027174271643161774, + "learning_rate": 0.00027186667240458617, + "loss": 3.19810152053833, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.028479408472776413, + "learning_rate": 0.0002716015671438419, + "loss": 3.2178421020507812, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.026820644736289978, + "learning_rate": 0.0002713364871802478, + "loss": 3.2634482383728027, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.026427917182445526, + "learning_rate": 0.0002710714327273741, + "loss": 3.2452914714813232, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.02679610252380371, + "learning_rate": 0.0002708064039987704, + "loss": 3.243825912475586, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.026115626096725464, + "learning_rate": 0.00027054140120796546, + "loss": 3.218787670135498, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.026686571538448334, + "learning_rate": 0.0002702764245684672, + "loss": 3.2785303592681885, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.02685859426856041, + "learning_rate": 0.0002700114742937627, + "loss": 3.2572226524353027, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.026968171820044518, + "learning_rate": 0.00026974655059731754, + "loss": 3.2625553607940674, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.025397246703505516, + "learning_rate": 0.0002694816536925759, + "loss": 3.262303352355957, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.025660471990704536, + "learning_rate": 0.00026921678379296057, + "loss": 3.269401788711548, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.02713891491293907, + "learning_rate": 0.0002689519411118722, + "loss": 3.2416529655456543, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.027591092512011528, + "learning_rate": 0.00026868712586269, + "loss": 3.2814316749572754, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.024462848901748657, + "learning_rate": 0.00026842233825877064, + "loss": 3.249725103378296, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.024868478998541832, + "learning_rate": 0.0002681575785134487, + "loss": 3.231574535369873, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.027015401050448418, + "learning_rate": 0.0002678928468400366, + "loss": 3.282627582550049, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.02678140625357628, + "learning_rate": 0.00026762814345182367, + "loss": 3.2208900451660156, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.026830527931451797, + "learning_rate": 0.0002673634685620767, + "loss": 3.260575294494629, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.02430890128016472, + "learning_rate": 0.0002670988223840393, + "loss": 3.2522997856140137, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.027047613635659218, + "learning_rate": 0.0002668342051309324, + "loss": 3.2721078395843506, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.026861369609832764, + "learning_rate": 0.00026656961701595306, + "loss": 3.2235946655273438, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.026650549843907356, + "learning_rate": 0.0002663050582522753, + "loss": 3.302743434906006, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.02551819011569023, + "learning_rate": 0.00026604052905304946, + "loss": 3.265763282775879, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.024843450635671616, + "learning_rate": 0.00026577602963140156, + "loss": 3.272397518157959, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.024423206225037575, + "learning_rate": 0.0002655115602004345, + "loss": 3.253495454788208, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.024180661886930466, + "learning_rate": 0.00026524712097322604, + "loss": 3.260438919067383, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.023554982617497444, + "learning_rate": 0.0002649827121628305, + "loss": 3.2342851161956787, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.023470086976885796, + "learning_rate": 0.00026471833398227694, + "loss": 3.251622438430786, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.024405231699347496, + "learning_rate": 0.0002644539866445703, + "loss": 3.2588815689086914, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.024147814139723778, + "learning_rate": 0.00026418967036269045, + "loss": 3.2723565101623535, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.023217910900712013, + "learning_rate": 0.0002639253853495921, + "loss": 3.2555792331695557, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.024921158328652382, + "learning_rate": 0.00026366113181820506, + "loss": 3.26884126663208, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.02455919235944748, + "learning_rate": 0.00026339690998143346, + "loss": 3.2674927711486816, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.024725932627916336, + "learning_rate": 0.00026313272005215637, + "loss": 3.2511110305786133, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.025769414380192757, + "learning_rate": 0.00026286856224322645, + "loss": 3.290651321411133, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.025928806513547897, + "learning_rate": 0.00026260443676747114, + "loss": 3.302396297454834, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.02499072253704071, + "learning_rate": 0.00026234034383769157, + "loss": 3.2813124656677246, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.025431277230381966, + "learning_rate": 0.0002620762836666625, + "loss": 3.234684705734253, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.026871860027313232, + "learning_rate": 0.00026181225646713257, + "loss": 3.2800331115722656, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.02578776702284813, + "learning_rate": 0.0002615482624518236, + "loss": 3.3117146492004395, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.026318378746509552, + "learning_rate": 0.0002612843018334309, + "loss": 3.257615089416504, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.025542527437210083, + "learning_rate": 0.00026102037482462256, + "loss": 3.2969868183135986, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.025723429396748543, + "learning_rate": 0.0002607564816380403, + "loss": 3.290259838104248, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.024588774889707565, + "learning_rate": 0.0002604926224862975, + "loss": 3.284090280532837, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.024619122967123985, + "learning_rate": 0.0002602287975819811, + "loss": 3.284879207611084, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.024441618472337723, + "learning_rate": 0.00025996500713765, + "loss": 3.283104658126831, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.02402760274708271, + "learning_rate": 0.0002597012513658352, + "loss": 3.3013617992401123, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.025028521195054054, + "learning_rate": 0.00025943753047904026, + "loss": 3.2892568111419678, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.029445001855492592, + "learning_rate": 0.0002591738446897401, + "loss": 3.2474446296691895, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.128176689147949, + "eval_runtime": 58.8588, + "eval_samples_per_second": 41.489, + "eval_steps_per_second": 1.308, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.0298826452344656, + "learning_rate": 0.00025891019421038177, + "loss": 3.138103485107422, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.0387548953294754, + "learning_rate": 0.00025864657925338345, + "loss": 3.161492347717285, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.03816070780158043, + "learning_rate": 0.00025838300003113514, + "loss": 3.1739673614501953, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.03409738838672638, + "learning_rate": 0.0002581194567559979, + "loss": 3.182478189468384, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.033135563135147095, + "learning_rate": 0.00025785594964030365, + "loss": 3.1720199584960938, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.03203624114394188, + "learning_rate": 0.0002575924788963554, + "loss": 3.1699862480163574, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.031237754970788956, + "learning_rate": 0.0002573290447364268, + "loss": 3.179363250732422, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.03342146426439285, + "learning_rate": 0.0002570656473727619, + "loss": 3.1822280883789062, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.03561089560389519, + "learning_rate": 0.0002568022870175752, + "loss": 3.1767094135284424, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.034710463136434555, + "learning_rate": 0.00025653896388305134, + "loss": 3.218817949295044, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.030870331451296806, + "learning_rate": 0.0002562756781813453, + "loss": 3.168602705001831, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.028004391118884087, + "learning_rate": 0.00025601243012458126, + "loss": 3.224155902862549, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.030953701585531235, + "learning_rate": 0.00025574921992485366, + "loss": 3.1789557933807373, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.030289432033896446, + "learning_rate": 0.000255486047794226, + "loss": 3.1890387535095215, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.030684132128953934, + "learning_rate": 0.0002552229139447316, + "loss": 3.195427417755127, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.03022196516394615, + "learning_rate": 0.00025495981858837235, + "loss": 3.1984965801239014, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.026935536414384842, + "learning_rate": 0.0002546967619371195, + "loss": 3.2014272212982178, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.02702922560274601, + "learning_rate": 0.0002544337442029133, + "loss": 3.205538034439087, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.026682600378990173, + "learning_rate": 0.00025417076559766194, + "loss": 3.1954591274261475, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.0273384228348732, + "learning_rate": 0.000253907826333243, + "loss": 3.1782338619232178, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.028587577864527702, + "learning_rate": 0.00025364492662150145, + "loss": 3.199565887451172, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.028848132118582726, + "learning_rate": 0.0002533820666742514, + "loss": 3.181786298751831, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.027535341680049896, + "learning_rate": 0.00025311924670327364, + "loss": 3.2013628482818604, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.02756025455892086, + "learning_rate": 0.00025285646692031804, + "loss": 3.1749587059020996, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.027358995750546455, + "learning_rate": 0.0002525937275371013, + "loss": 3.161881923675537, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.025918981060385704, + "learning_rate": 0.00025233102876530785, + "loss": 3.225224256515503, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.026479976251721382, + "learning_rate": 0.0002520683708165894, + "loss": 3.2129316329956055, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.02685079537332058, + "learning_rate": 0.0002518057539025646, + "loss": 3.1651554107666016, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.027378706261515617, + "learning_rate": 0.00025154317823481917, + "loss": 3.193108558654785, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.027240974828600883, + "learning_rate": 0.0002512806440249056, + "loss": 3.190605640411377, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.028065402060747147, + "learning_rate": 0.00025101815148434296, + "loss": 3.2068259716033936, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.028329014778137207, + "learning_rate": 0.00025075570082461684, + "loss": 3.2316060066223145, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.028920065611600876, + "learning_rate": 0.0002504932922571789, + "loss": 3.1933932304382324, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.028299354016780853, + "learning_rate": 0.00025023092599344706, + "loss": 3.192711353302002, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.025996020063757896, + "learning_rate": 0.0002499686022448049, + "loss": 3.243171453475952, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.0264595914632082, + "learning_rate": 0.00024970632122260216, + "loss": 3.199758529663086, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.02704528719186783, + "learning_rate": 0.0002494440831381538, + "loss": 3.2390713691711426, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.026571933180093765, + "learning_rate": 0.00024918188820274016, + "loss": 3.219440460205078, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.025562983006238937, + "learning_rate": 0.0002489197366276073, + "loss": 3.2538228034973145, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.024683184921741486, + "learning_rate": 0.0002486576286239657, + "loss": 3.2372591495513916, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.025656528770923615, + "learning_rate": 0.0002483955644029913, + "loss": 3.250558853149414, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.024807540699839592, + "learning_rate": 0.0002481335441758243, + "loss": 3.1856563091278076, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.026149194687604904, + "learning_rate": 0.0002478715681535699, + "loss": 3.172186851501465, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.02617705799639225, + "learning_rate": 0.00024760963654729717, + "loss": 3.250445604324341, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.026542512699961662, + "learning_rate": 0.0002473477495680397, + "loss": 3.220273494720459, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.024963509291410446, + "learning_rate": 0.00024708590742679546, + "loss": 3.2330946922302246, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.027144264429807663, + "learning_rate": 0.00024682411033452546, + "loss": 3.1860971450805664, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.02665683813393116, + "learning_rate": 0.0002465623585021553, + "loss": 3.257209062576294, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.024872658774256706, + "learning_rate": 0.0002463006521405733, + "loss": 3.2266182899475098, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.025235185399651527, + "learning_rate": 0.0002460389914606319, + "loss": 3.2134504318237305, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.025029495358467102, + "learning_rate": 0.00024577737667314615, + "loss": 3.2150087356567383, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.025630896911025047, + "learning_rate": 0.00024551580798889446, + "loss": 3.282879114151001, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.02722341939806938, + "learning_rate": 0.00024525428561861804, + "loss": 3.2366206645965576, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.02522372268140316, + "learning_rate": 0.0002449928097730207, + "loss": 3.2840523719787598, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.02494790591299534, + "learning_rate": 0.00024473138066276876, + "loss": 3.2405319213867188, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.026662737131118774, + "learning_rate": 0.0002444699984984909, + "loss": 3.239692211151123, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.026082608848810196, + "learning_rate": 0.0002442086634907782, + "loss": 3.2156195640563965, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.02423974685370922, + "learning_rate": 0.00024394737585018328, + "loss": 3.1863760948181152, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.025226688012480736, + "learning_rate": 0.00024368613578722113, + "loss": 3.20874285697937, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.02428225241601467, + "learning_rate": 0.00024342494351236799, + "loss": 3.2209601402282715, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.023819707334041595, + "learning_rate": 0.0002431637992360618, + "loss": 3.2564258575439453, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.025661710649728775, + "learning_rate": 0.00024290270316870184, + "loss": 3.244133949279785, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.02423049882054329, + "learning_rate": 0.0002426416555206484, + "loss": 3.2376630306243896, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.02382473088800907, + "learning_rate": 0.000242380656502223, + "loss": 3.2951290607452393, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.02488875202834606, + "learning_rate": 0.00024211970632370756, + "loss": 3.2848310470581055, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.02575223706662655, + "learning_rate": 0.0002418588051953453, + "loss": 3.2574503421783447, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.02462848648428917, + "learning_rate": 0.0002415979533273392, + "loss": 3.2464442253112793, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.024912388995289803, + "learning_rate": 0.0002413371509298531, + "loss": 3.2459192276000977, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.025020884349942207, + "learning_rate": 0.00024107639821301078, + "loss": 3.213622808456421, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.026330886408686638, + "learning_rate": 0.00024081569538689587, + "loss": 3.2305431365966797, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.02708602324128151, + "learning_rate": 0.0002405550426615521, + "loss": 3.2487688064575195, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.030908891931176186, + "learning_rate": 0.00024029444024698248, + "loss": 3.2110159397125244, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.146673202514648, + "eval_runtime": 58.7379, + "eval_samples_per_second": 41.574, + "eval_steps_per_second": 1.311, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.02919941022992134, + "learning_rate": 0.00024003388835314981, + "loss": 3.1533267498016357, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.03379782289266586, + "learning_rate": 0.00023977338718997595, + "loss": 3.1436538696289062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.033698830753564835, + "learning_rate": 0.00023951293696734195, + "loss": 3.15297532081604, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.03077736124396324, + "learning_rate": 0.00023925253789508806, + "loss": 3.128848075866699, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.031112324446439743, + "learning_rate": 0.00023899219018301286, + "loss": 3.1461119651794434, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.03346852958202362, + "learning_rate": 0.00023873189404087405, + "loss": 3.113943099975586, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.03307225927710533, + "learning_rate": 0.00023847164967838726, + "loss": 3.147498607635498, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.030230754986405373, + "learning_rate": 0.00023821145730522692, + "loss": 3.1162924766540527, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.02982981689274311, + "learning_rate": 0.0002379513171310251, + "loss": 3.144481658935547, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.031735632568597794, + "learning_rate": 0.00023769122936537207, + "loss": 3.133862018585205, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.0312766470015049, + "learning_rate": 0.00023743119421781614, + "loss": 3.110287666320801, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.029902562499046326, + "learning_rate": 0.00023717121189786256, + "loss": 3.1384904384613037, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.0306564774364233, + "learning_rate": 0.00023691128261497463, + "loss": 3.1138200759887695, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.03084711916744709, + "learning_rate": 0.0002366514065785725, + "loss": 3.155240535736084, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.03183220326900482, + "learning_rate": 0.00023639158399803382, + "loss": 3.1593515872955322, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.031826410442590714, + "learning_rate": 0.0002361318150826927, + "loss": 3.1517653465270996, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.030141707509756088, + "learning_rate": 0.0002358721000418404, + "loss": 3.1291589736938477, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.03158571943640709, + "learning_rate": 0.00023561243908472472, + "loss": 3.178101062774658, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.03510458394885063, + "learning_rate": 0.00023535283242054954, + "loss": 3.138767719268799, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.03660859912633896, + "learning_rate": 0.00023509328025847565, + "loss": 3.1708762645721436, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.034004807472229004, + "learning_rate": 0.00023483378280761907, + "loss": 3.1404073238372803, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.03077496401965618, + "learning_rate": 0.0002345743402770527, + "loss": 3.159081220626831, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.030384650453925133, + "learning_rate": 0.0002343149528758042, + "loss": 3.1872847080230713, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.02957277186214924, + "learning_rate": 0.00023405562081285766, + "loss": 3.165973424911499, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.029529055580496788, + "learning_rate": 0.0002337963442971522, + "loss": 3.123840808868408, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.027776138857007027, + "learning_rate": 0.00023353712353758212, + "loss": 3.1705269813537598, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.028943249955773354, + "learning_rate": 0.00023327795874299704, + "loss": 3.1181654930114746, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.028178401291370392, + "learning_rate": 0.0002330188501222012, + "loss": 3.201678991317749, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.02865665778517723, + "learning_rate": 0.0002327597978839538, + "loss": 3.18479061126709, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.029173649847507477, + "learning_rate": 0.00023250080223696852, + "loss": 3.1844916343688965, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.0280170775949955, + "learning_rate": 0.00023224186338991352, + "loss": 3.187105417251587, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.029103878885507584, + "learning_rate": 0.0002319829815514112, + "loss": 3.183281660079956, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.029045486822724342, + "learning_rate": 0.00023172415693003789, + "loss": 3.1848196983337402, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.028364380821585655, + "learning_rate": 0.000231465389734324, + "loss": 3.197348117828369, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.029651399701833725, + "learning_rate": 0.00023120668017275346, + "loss": 3.1429920196533203, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.0281438659876585, + "learning_rate": 0.00023094802845376403, + "loss": 3.1916513442993164, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.029437677934765816, + "learning_rate": 0.00023068943478574662, + "loss": 3.172391176223755, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.03433044254779816, + "learning_rate": 0.00023043089937704541, + "loss": 3.2296864986419678, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.03259492665529251, + "learning_rate": 0.00023017242243595796, + "loss": 3.1962363719940186, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.0299630519002676, + "learning_rate": 0.00022991400417073427, + "loss": 3.169673442840576, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.028495769947767258, + "learning_rate": 0.00022965564478957739, + "loss": 3.1855409145355225, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.0290420800447464, + "learning_rate": 0.0002293973445006427, + "loss": 3.23667573928833, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.03340664133429527, + "learning_rate": 0.00022913910351203816, + "loss": 3.1873788833618164, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.03253626823425293, + "learning_rate": 0.00022888092203182377, + "loss": 3.204796552658081, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.028142794966697693, + "learning_rate": 0.00022862280026801182, + "loss": 3.193922281265259, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.028805622830986977, + "learning_rate": 0.00022836473842856623, + "loss": 3.1863901615142822, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.030033612623810768, + "learning_rate": 0.00022810673672140272, + "loss": 3.198482036590576, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.028154220432043076, + "learning_rate": 0.0002278487953543889, + "loss": 3.215069055557251, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.028098132461309433, + "learning_rate": 0.00022759091453534308, + "loss": 3.2197628021240234, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.028679659590125084, + "learning_rate": 0.00022733309447203553, + "loss": 3.2062196731567383, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.029104258865118027, + "learning_rate": 0.00022707533537218695, + "loss": 3.2182908058166504, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.029629778116941452, + "learning_rate": 0.00022681763744346938, + "loss": 3.2165639400482178, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.028304366394877434, + "learning_rate": 0.00022656000089350524, + "loss": 3.1783785820007324, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.026900488883256912, + "learning_rate": 0.00022630242592986766, + "loss": 3.1876602172851562, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.027047287672758102, + "learning_rate": 0.0002260449127600803, + "loss": 3.1879465579986572, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.027424108237028122, + "learning_rate": 0.0002257874615916166, + "loss": 3.2407851219177246, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.029319508001208305, + "learning_rate": 0.00022553007263190067, + "loss": 3.1980130672454834, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.027114372700452805, + "learning_rate": 0.00022527274608830574, + "loss": 3.211357593536377, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.02579840086400509, + "learning_rate": 0.0002250154821681555, + "loss": 3.2067861557006836, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.02767331339418888, + "learning_rate": 0.0002247582810787227, + "loss": 3.198587417602539, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.026645265519618988, + "learning_rate": 0.00022450114302722957, + "loss": 3.177441120147705, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.026181906461715698, + "learning_rate": 0.0002242440682208477, + "loss": 3.226815700531006, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.027218692004680634, + "learning_rate": 0.00022398705686669748, + "loss": 3.1731510162353516, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.027551919221878052, + "learning_rate": 0.0002237301091718485, + "loss": 3.2268621921539307, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.025782009586691856, + "learning_rate": 0.00022347322534331865, + "loss": 3.2052621841430664, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.024231620132923126, + "learning_rate": 0.00022321640558807488, + "loss": 3.201770544052124, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.02474607340991497, + "learning_rate": 0.00022295965011303188, + "loss": 3.1935834884643555, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.026299327611923218, + "learning_rate": 0.00022270295912505317, + "loss": 3.223275899887085, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.02439914271235466, + "learning_rate": 0.00022244633283095, + "loss": 3.2513954639434814, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.02650653012096882, + "learning_rate": 0.0002221897714374814, + "loss": 3.1914851665496826, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.024847298860549927, + "learning_rate": 0.00022193327515135445, + "loss": 3.2173104286193848, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.028475770726799965, + "learning_rate": 0.00022167684417922337, + "loss": 3.2184863090515137, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.168707847595215, + "eval_runtime": 57.4035, + "eval_samples_per_second": 42.541, + "eval_steps_per_second": 1.341, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.02837727963924408, + "learning_rate": 0.00022142047872769014, + "loss": 3.119239330291748, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.03562312573194504, + "learning_rate": 0.00022116417900330365, + "loss": 3.092813014984131, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.0344194732606411, + "learning_rate": 0.00022090794521255997, + "loss": 3.0802855491638184, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.03188958019018173, + "learning_rate": 0.00022065177756190214, + "loss": 3.1177961826324463, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.032181013375520706, + "learning_rate": 0.00022039567625771968, + "loss": 3.1343281269073486, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.031085725873708725, + "learning_rate": 0.0002201396415063489, + "loss": 3.118828296661377, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.031995322555303574, + "learning_rate": 0.00021988367351407227, + "loss": 3.110109329223633, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.02949182130396366, + "learning_rate": 0.00021962777248711872, + "loss": 3.0972743034362793, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.029299668967723846, + "learning_rate": 0.00021937193863166292, + "loss": 3.1257238388061523, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.030558617785573006, + "learning_rate": 0.00021911617215382564, + "loss": 3.1216022968292236, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.03181575983762741, + "learning_rate": 0.00021886047325967348, + "loss": 3.093721866607666, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.030934303998947144, + "learning_rate": 0.00021860484215521805, + "loss": 3.0910396575927734, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.030235569924116135, + "learning_rate": 0.00021834927904641708, + "loss": 3.133371114730835, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.03350163996219635, + "learning_rate": 0.00021809378413917296, + "loss": 3.1178882122039795, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.03468719869852066, + "learning_rate": 0.00021783835763933344, + "loss": 3.1222469806671143, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.034203238785266876, + "learning_rate": 0.0002175829997526909, + "loss": 3.12453031539917, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.03321516141295433, + "learning_rate": 0.0002173277106849826, + "loss": 3.115428924560547, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.03298434242606163, + "learning_rate": 0.0002170724906418905, + "loss": 3.127288341522217, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.029987169429659843, + "learning_rate": 0.00021681733982904046, + "loss": 3.1194515228271484, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.031271349638700485, + "learning_rate": 0.00021656225845200332, + "loss": 3.1284165382385254, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.032201699912548065, + "learning_rate": 0.0002163072467162931, + "loss": 3.1912736892700195, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.0299740731716156, + "learning_rate": 0.00021605230482736845, + "loss": 3.1233785152435303, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.029613154008984566, + "learning_rate": 0.00021579743299063122, + "loss": 3.1535067558288574, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.03303765505552292, + "learning_rate": 0.00021554263141142726, + "loss": 3.1550910472869873, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.032644059509038925, + "learning_rate": 0.00021528790029504548, + "loss": 3.1367573738098145, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.031695179641246796, + "learning_rate": 0.00021503323984671814, + "loss": 3.1350858211517334, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.030685821548104286, + "learning_rate": 0.00021477865027162067, + "loss": 3.121593952178955, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.03091946430504322, + "learning_rate": 0.0002145241317748711, + "loss": 3.113060235977173, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.0282905176281929, + "learning_rate": 0.0002142696845615306, + "loss": 3.184551239013672, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.030020160600543022, + "learning_rate": 0.0002140153088366024, + "loss": 3.1673004627227783, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.031114062294363976, + "learning_rate": 0.00021376100480503276, + "loss": 3.162588596343994, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.03007666766643524, + "learning_rate": 0.00021350677267170973, + "loss": 3.132836103439331, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.029013793915510178, + "learning_rate": 0.00021325261264146342, + "loss": 3.1552720069885254, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.02784561738371849, + "learning_rate": 0.00021299852491906616, + "loss": 3.189272165298462, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.02759280428290367, + "learning_rate": 0.00021274450970923167, + "loss": 3.141814708709717, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.02699868381023407, + "learning_rate": 0.0002124905672166156, + "loss": 3.1127257347106934, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.02782190404832363, + "learning_rate": 0.0002122366976458146, + "loss": 3.193972587585449, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.027298318222165108, + "learning_rate": 0.00021198290120136702, + "loss": 3.1393446922302246, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.027321161702275276, + "learning_rate": 0.0002117291780877519, + "loss": 3.0926666259765625, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.026399722322821617, + "learning_rate": 0.00021147552850938941, + "loss": 3.1578259468078613, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.02770969271659851, + "learning_rate": 0.0002112219526706406, + "loss": 3.1816914081573486, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.028344258666038513, + "learning_rate": 0.00021096845077580675, + "loss": 3.1860158443450928, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.028091782703995705, + "learning_rate": 0.0002107150230291299, + "loss": 3.1585092544555664, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.026564039289951324, + "learning_rate": 0.00021046166963479204, + "loss": 3.131692409515381, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.026349127292633057, + "learning_rate": 0.00021020839079691553, + "loss": 3.160431385040283, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.030081138014793396, + "learning_rate": 0.00020995518671956256, + "loss": 3.1647543907165527, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.029250919818878174, + "learning_rate": 0.00020970205760673493, + "loss": 3.1522226333618164, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.0282477755099535, + "learning_rate": 0.00020944900366237453, + "loss": 3.139249324798584, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.0284534003585577, + "learning_rate": 0.00020919602509036195, + "loss": 3.154621124267578, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.02677394263446331, + "learning_rate": 0.0002089431220945178, + "loss": 3.1375081539154053, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.026511570438742638, + "learning_rate": 0.00020869029487860126, + "loss": 3.1355676651000977, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.02766270563006401, + "learning_rate": 0.00020843754364631068, + "loss": 3.143874168395996, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.027013977989554405, + "learning_rate": 0.0002081848686012832, + "loss": 3.1602487564086914, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.026629449799656868, + "learning_rate": 0.00020793226994709446, + "loss": 3.1549456119537354, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.0272684209048748, + "learning_rate": 0.0002076797478872588, + "loss": 3.1739020347595215, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.027645915746688843, + "learning_rate": 0.00020742730262522844, + "loss": 3.172959327697754, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.025413552299141884, + "learning_rate": 0.00020717493436439433, + "loss": 3.1649041175842285, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.028406929224729538, + "learning_rate": 0.00020692264330808455, + "loss": 3.1821887493133545, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.027753977105021477, + "learning_rate": 0.00020667042965956587, + "loss": 3.163238048553467, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.02693209983408451, + "learning_rate": 0.00020641829362204195, + "loss": 3.1921019554138184, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.02671878971159458, + "learning_rate": 0.00020616623539865446, + "loss": 3.159438133239746, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.027099428698420525, + "learning_rate": 0.00020591425519248213, + "loss": 3.1806793212890625, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.025098824873566628, + "learning_rate": 0.00020566235320654069, + "loss": 3.158310651779175, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.02655479498207569, + "learning_rate": 0.00020541052964378323, + "loss": 3.1927309036254883, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.025836164131760597, + "learning_rate": 0.00020515878470709928, + "loss": 3.1578540802001953, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.02657160349190235, + "learning_rate": 0.00020490711859931545, + "loss": 3.1658592224121094, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.027223270386457443, + "learning_rate": 0.00020465553152319418, + "loss": 3.1760241985321045, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.025695348158478737, + "learning_rate": 0.00020440402368143492, + "loss": 3.202389717102051, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.025819813832640648, + "learning_rate": 0.00020415259527667299, + "loss": 3.2111103534698486, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.02621106244623661, + "learning_rate": 0.0002039012465114796, + "loss": 3.1560299396514893, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.026585502550005913, + "learning_rate": 0.00020364997758836193, + "loss": 3.176602840423584, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.03038652054965496, + "learning_rate": 0.00020339878870976287, + "loss": 3.1799943447113037, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.193845748901367, + "eval_runtime": 57.9911, + "eval_samples_per_second": 42.11, + "eval_steps_per_second": 1.328, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.03135736659169197, + "learning_rate": 0.00020314768007806077, + "loss": 3.070986270904541, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.038986921310424805, + "learning_rate": 0.00020289665189556923, + "loss": 3.0708835124969482, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.035115309059619904, + "learning_rate": 0.00020264570436453707, + "loss": 3.1070001125335693, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.032868221402168274, + "learning_rate": 0.00020239483768714843, + "loss": 3.082258701324463, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.03703184798359871, + "learning_rate": 0.0002021440520655219, + "loss": 3.057744264602661, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.03831101953983307, + "learning_rate": 0.00020189334770171098, + "loss": 3.0789475440979004, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.036301176995038986, + "learning_rate": 0.00020164272479770352, + "loss": 3.0565457344055176, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.03166409209370613, + "learning_rate": 0.00020139218355542203, + "loss": 3.0574910640716553, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.02984458953142166, + "learning_rate": 0.000201141724176723, + "loss": 3.080018997192383, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.034001126885414124, + "learning_rate": 0.0002008913468633969, + "loss": 3.0813302993774414, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.03814122825860977, + "learning_rate": 0.00020064105181716847, + "loss": 3.0579185485839844, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.03870106488466263, + "learning_rate": 0.00020039083923969564, + "loss": 3.0857529640197754, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.04005642607808113, + "learning_rate": 0.0002001407093325704, + "loss": 3.0880727767944336, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.03760755434632301, + "learning_rate": 0.0001998906622973177, + "loss": 3.0795767307281494, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.03260907158255577, + "learning_rate": 0.0001996406983353961, + "loss": 3.1006722450256348, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.029866183176636696, + "learning_rate": 0.0001993908176481969, + "loss": 3.0850343704223633, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.030988436192274094, + "learning_rate": 0.0001991410204370445, + "loss": 3.0743892192840576, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.03698134422302246, + "learning_rate": 0.00019889130690319608, + "loss": 3.056002378463745, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.03696191683411598, + "learning_rate": 0.00019864167724784117, + "loss": 3.069124698638916, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.03453022986650467, + "learning_rate": 0.00019839213167210213, + "loss": 3.1017374992370605, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.032458193600177765, + "learning_rate": 0.00019814267037703294, + "loss": 3.1166419982910156, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.02892368845641613, + "learning_rate": 0.0001978932935636205, + "loss": 3.0863287448883057, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.030625877901911736, + "learning_rate": 0.00019764400143278276, + "loss": 3.0938801765441895, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.030280394479632378, + "learning_rate": 0.00019739479418537018, + "loss": 3.086432933807373, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.029181072488427162, + "learning_rate": 0.00019714567202216453, + "loss": 3.0977859497070312, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.028039127588272095, + "learning_rate": 0.00019689663514387896, + "loss": 3.112367868423462, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.02910272404551506, + "learning_rate": 0.00019664768375115798, + "loss": 3.1253533363342285, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.03086872771382332, + "learning_rate": 0.00019639881804457723, + "loss": 3.103689670562744, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.02883610874414444, + "learning_rate": 0.00019615003822464342, + "loss": 3.111599922180176, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.028490116819739342, + "learning_rate": 0.00019590134449179371, + "loss": 3.1410303115844727, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.028714407235383987, + "learning_rate": 0.00019565273704639648, + "loss": 3.133023738861084, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.029553238302469254, + "learning_rate": 0.00019540421608874994, + "loss": 3.107003688812256, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.02997349575161934, + "learning_rate": 0.0001951557818190831, + "loss": 3.135680675506592, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.029087591916322708, + "learning_rate": 0.00019490743443755504, + "loss": 3.1103882789611816, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.029947761446237564, + "learning_rate": 0.0001946591741442546, + "loss": 3.0961623191833496, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.02950555644929409, + "learning_rate": 0.00019441100113920076, + "loss": 3.1162495613098145, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.029569562524557114, + "learning_rate": 0.00019416291562234192, + "loss": 3.096372127532959, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.031366728246212006, + "learning_rate": 0.00019391491779355608, + "loss": 3.120089530944824, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.0313473716378212, + "learning_rate": 0.00019366700785265065, + "loss": 3.100785732269287, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.028021851554512978, + "learning_rate": 0.0001934191859993622, + "loss": 3.0940558910369873, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.027477117255330086, + "learning_rate": 0.00019317145243335642, + "loss": 3.1210484504699707, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.029690749943256378, + "learning_rate": 0.0001929238073542276, + "loss": 3.1350059509277344, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.029659733176231384, + "learning_rate": 0.00019267625096149903, + "loss": 3.1220788955688477, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.028260033577680588, + "learning_rate": 0.00019242878345462227, + "loss": 3.1000189781188965, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.027554461732506752, + "learning_rate": 0.00019218140503297765, + "loss": 3.1044349670410156, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.027929967269301414, + "learning_rate": 0.00019193411589587332, + "loss": 3.1353745460510254, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.029309319332242012, + "learning_rate": 0.00019168691624254564, + "loss": 3.1578292846679688, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.029183287173509598, + "learning_rate": 0.00019143980627215915, + "loss": 3.1217150688171387, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.02707512304186821, + "learning_rate": 0.00019119278618380555, + "loss": 3.119617223739624, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.03006536327302456, + "learning_rate": 0.00019094585617650476, + "loss": 3.1481480598449707, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.03026784397661686, + "learning_rate": 0.00019069901644920366, + "loss": 3.142770767211914, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.02827967144548893, + "learning_rate": 0.00019045226720077667, + "loss": 3.1582541465759277, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.028892774134874344, + "learning_rate": 0.0001902056086300251, + "loss": 3.1530709266662598, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.03023783676326275, + "learning_rate": 0.0001899590409356773, + "loss": 3.1102521419525146, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.027951840311288834, + "learning_rate": 0.00018971256431638854, + "loss": 3.139113426208496, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.02807226963341236, + "learning_rate": 0.00018946617897074041, + "loss": 3.149428367614746, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.029721451923251152, + "learning_rate": 0.0001892198850972413, + "loss": 3.120518207550049, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.02867680788040161, + "learning_rate": 0.00018897368289432553, + "loss": 3.1109931468963623, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.02663845382630825, + "learning_rate": 0.00018872757256035403, + "loss": 3.116053581237793, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.028366081416606903, + "learning_rate": 0.00018848155429361323, + "loss": 3.1444268226623535, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.028931576758623123, + "learning_rate": 0.00018823562829231567, + "loss": 3.163870096206665, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.02664567157626152, + "learning_rate": 0.0001879897947545996, + "loss": 3.1349146366119385, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.028031354770064354, + "learning_rate": 0.00018774405387852848, + "loss": 3.1638219356536865, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.029421644285321236, + "learning_rate": 0.00018749840586209144, + "loss": 3.1601340770721436, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.027983693405985832, + "learning_rate": 0.00018725285090320245, + "loss": 3.1636788845062256, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.02719871513545513, + "learning_rate": 0.00018700738919970103, + "loss": 3.1246869564056396, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.03030397929251194, + "learning_rate": 0.00018676202094935085, + "loss": 3.1749579906463623, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.029340893030166626, + "learning_rate": 0.000186516746349841, + "loss": 3.1751790046691895, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.027624106034636497, + "learning_rate": 0.0001862715655987847, + "loss": 3.178704023361206, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.02830546908080578, + "learning_rate": 0.00018602647889371957, + "loss": 3.177117347717285, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.03168068826198578, + "learning_rate": 0.00018578148643210764, + "loss": 3.125427722930908, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.033327918499708176, + "learning_rate": 0.00018553658841133483, + "loss": 3.13071870803833, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.218388080596924, + "eval_runtime": 57.611, + "eval_samples_per_second": 42.388, + "eval_steps_per_second": 1.337, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.03336749225854874, + "learning_rate": 0.00018529178502871117, + "loss": 3.064338445663452, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.038479577749967575, + "learning_rate": 0.0001850470764814702, + "loss": 3.0478687286376953, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.032562077045440674, + "learning_rate": 0.0001848024629667691, + "loss": 3.019803762435913, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.0331728458404541, + "learning_rate": 0.00018455794468168887, + "loss": 3.0599141120910645, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.03361169621348381, + "learning_rate": 0.00018431352182323322, + "loss": 3.0315823554992676, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.032973092049360275, + "learning_rate": 0.0001840691945883293, + "loss": 3.025928258895874, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.030388766899704933, + "learning_rate": 0.00018382496317382715, + "loss": 3.070671558380127, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.030899515375494957, + "learning_rate": 0.0001835808277764997, + "loss": 3.012938976287842, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.0335412323474884, + "learning_rate": 0.00018333678859304222, + "loss": 3.0156331062316895, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.035731345415115356, + "learning_rate": 0.00018309284582007274, + "loss": 3.0578935146331787, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.03379053995013237, + "learning_rate": 0.0001828489996541318, + "loss": 3.048701286315918, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.031242679804563522, + "learning_rate": 0.00018260525029168147, + "loss": 3.0832977294921875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.030997611582279205, + "learning_rate": 0.00018236159792910648, + "loss": 3.0663583278656006, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.03397331386804581, + "learning_rate": 0.00018211804276271304, + "loss": 3.0115303993225098, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.03349050134420395, + "learning_rate": 0.00018187458498872914, + "loss": 3.02864933013916, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.031079065054655075, + "learning_rate": 0.00018163122480330433, + "loss": 3.053401470184326, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.029346197843551636, + "learning_rate": 0.00018138796240250955, + "loss": 3.0700926780700684, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.030784208327531815, + "learning_rate": 0.00018114479798233686, + "loss": 3.066525936126709, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.03368140384554863, + "learning_rate": 0.00018090173173869939, + "loss": 3.0956242084503174, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.03356792405247688, + "learning_rate": 0.00018065876386743143, + "loss": 3.0394320487976074, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.029909325763583183, + "learning_rate": 0.00018041589456428754, + "loss": 3.031525135040283, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.030711578205227852, + "learning_rate": 0.0001801731240249434, + "loss": 3.075474977493286, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.03452072665095329, + "learning_rate": 0.00017993045244499463, + "loss": 3.0228147506713867, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.03235075995326042, + "learning_rate": 0.00017968788001995742, + "loss": 3.0664501190185547, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.029671598225831985, + "learning_rate": 0.000179445406945268, + "loss": 3.055302143096924, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.033868469297885895, + "learning_rate": 0.0001792030334162825, + "loss": 3.053016185760498, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.0329158715903759, + "learning_rate": 0.00017896075962827706, + "loss": 3.1017332077026367, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.031983695924282074, + "learning_rate": 0.00017871858577644712, + "loss": 3.0706472396850586, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.02963697351515293, + "learning_rate": 0.0001784765120559078, + "loss": 3.0711426734924316, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.03162391856312752, + "learning_rate": 0.00017823453866169355, + "loss": 3.044867515563965, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.030773654580116272, + "learning_rate": 0.00017799266578875808, + "loss": 3.0963196754455566, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.031220417469739914, + "learning_rate": 0.00017775089363197371, + "loss": 3.0892245769500732, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.030928561463952065, + "learning_rate": 0.00017750922238613198, + "loss": 3.1181323528289795, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.03192073106765747, + "learning_rate": 0.00017726765224594342, + "loss": 3.066584348678589, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.03075495734810829, + "learning_rate": 0.0001770261834060363, + "loss": 3.0436511039733887, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.02895325981080532, + "learning_rate": 0.00017678481606095786, + "loss": 3.123373031616211, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.029738714918494225, + "learning_rate": 0.0001765435504051734, + "loss": 3.1049060821533203, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.031937118619680405, + "learning_rate": 0.0001763023866330664, + "loss": 3.099111318588257, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.029886579141020775, + "learning_rate": 0.00017606132493893793, + "loss": 3.0970301628112793, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.029569782316684723, + "learning_rate": 0.00017582036551700724, + "loss": 3.062499761581421, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.02924283780157566, + "learning_rate": 0.00017557950856141085, + "loss": 3.0618538856506348, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.030955800786614418, + "learning_rate": 0.00017533875426620286, + "loss": 3.0745606422424316, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.03145027533173561, + "learning_rate": 0.00017509810282535497, + "loss": 3.122938632965088, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.028926467522978783, + "learning_rate": 0.00017485755443275513, + "loss": 3.0808773040771484, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.029592076316475868, + "learning_rate": 0.0001746171092822094, + "loss": 3.1167049407958984, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.028070243075489998, + "learning_rate": 0.00017437676756743974, + "loss": 3.0781636238098145, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.028642063960433006, + "learning_rate": 0.00017413652948208524, + "loss": 3.093937397003174, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.0290757454931736, + "learning_rate": 0.00017389639521970136, + "loss": 3.1079559326171875, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.027867283672094345, + "learning_rate": 0.00017365636497376004, + "loss": 3.0885753631591797, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.02896495908498764, + "learning_rate": 0.00017341643893764927, + "loss": 3.0889246463775635, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.02777806483209133, + "learning_rate": 0.000173176617304673, + "loss": 3.084439277648926, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.02904277853667736, + "learning_rate": 0.00017293690026805124, + "loss": 3.125246047973633, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.029010627418756485, + "learning_rate": 0.00017269728802091975, + "loss": 3.111422061920166, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.028178894892334938, + "learning_rate": 0.00017245778075632975, + "loss": 3.115067481994629, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.02763799950480461, + "learning_rate": 0.00017221837866724792, + "loss": 3.1228506565093994, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.028052521869540215, + "learning_rate": 0.00017197908194655615, + "loss": 3.1150546073913574, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.029438253492116928, + "learning_rate": 0.00017173989078705175, + "loss": 3.0921788215637207, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.028195543214678764, + "learning_rate": 0.0001715008053814463, + "loss": 3.1408450603485107, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.02937602624297142, + "learning_rate": 0.00017126182592236682, + "loss": 3.111612319946289, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.028430236503481865, + "learning_rate": 0.00017102295260235472, + "loss": 3.093441963195801, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.027624569833278656, + "learning_rate": 0.0001707841856138659, + "loss": 3.1184372901916504, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.029797039926052094, + "learning_rate": 0.00017054552514927061, + "loss": 3.1074559688568115, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.028728218749165535, + "learning_rate": 0.00017030697140085334, + "loss": 3.161461114883423, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.02704412117600441, + "learning_rate": 0.0001700685245608126, + "loss": 3.13325834274292, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.028067929670214653, + "learning_rate": 0.00016983018482126047, + "loss": 3.093172550201416, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.028848597779870033, + "learning_rate": 0.00016959195237422305, + "loss": 3.1461825370788574, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.02956785447895527, + "learning_rate": 0.00016935382741164, + "loss": 3.141045570373535, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.028214773163199425, + "learning_rate": 0.00016911581012536424, + "loss": 3.136033773422241, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.029155246913433075, + "learning_rate": 0.0001688779007071622, + "loss": 3.0986382961273193, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.02968468703329563, + "learning_rate": 0.00016864009934871275, + "loss": 3.11630916595459, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.028356682509183884, + "learning_rate": 0.0001684024062416087, + "loss": 3.1134822368621826, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.03301718086004257, + "learning_rate": 0.0001681648215773546, + "loss": 3.118591785430908, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.246211528778076, + "eval_runtime": 57.5155, + "eval_samples_per_second": 42.458, + "eval_steps_per_second": 1.339, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.033637888729572296, + "learning_rate": 0.0001679273455473684, + "loss": 3.0192506313323975, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.04057888314127922, + "learning_rate": 0.0001676899783429802, + "loss": 2.977440595626831, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.03579854592680931, + "learning_rate": 0.00016745272015543242, + "loss": 2.9902219772338867, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.03131850063800812, + "learning_rate": 0.0001672155711758799, + "loss": 2.9988656044006348, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.03262636810541153, + "learning_rate": 0.00016697853159538896, + "loss": 3.037747621536255, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.03260403499007225, + "learning_rate": 0.0001667416016049383, + "loss": 2.987508773803711, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.033165134489536285, + "learning_rate": 0.0001665047813954181, + "loss": 3.0511183738708496, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.03303489834070206, + "learning_rate": 0.00016626807115763023, + "loss": 3.0083041191101074, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.03185006603598595, + "learning_rate": 0.00016603147108228754, + "loss": 2.9966228008270264, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.03135651350021362, + "learning_rate": 0.00016579498136001485, + "loss": 3.036759376525879, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.03216925635933876, + "learning_rate": 0.00016555860218134738, + "loss": 3.0304150581359863, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.03231671452522278, + "learning_rate": 0.00016532233373673164, + "loss": 3.0300474166870117, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.03263348713517189, + "learning_rate": 0.00016508617621652488, + "loss": 3.0298163890838623, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.032594650983810425, + "learning_rate": 0.00016485012981099488, + "loss": 2.98789119720459, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.03406650573015213, + "learning_rate": 0.0001646141947103201, + "loss": 3.0102760791778564, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.03379780799150467, + "learning_rate": 0.0001643783711045889, + "loss": 3.0564732551574707, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.03172587603330612, + "learning_rate": 0.0001641426591838005, + "loss": 3.0410943031311035, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.02938298135995865, + "learning_rate": 0.00016390705913786344, + "loss": 3.0517544746398926, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.03206252679228783, + "learning_rate": 0.00016367157115659646, + "loss": 3.0352776050567627, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.033386752009391785, + "learning_rate": 0.00016343619542972808, + "loss": 3.020535945892334, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.033984702080488205, + "learning_rate": 0.0001632009321468962, + "loss": 3.0552730560302734, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.0329931266605854, + "learning_rate": 0.00016296578149764833, + "loss": 3.0075573921203613, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.032101184129714966, + "learning_rate": 0.00016273074367144087, + "loss": 3.046388864517212, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.032873060554265976, + "learning_rate": 0.00016249581885763968, + "loss": 3.0340287685394287, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.030672110617160797, + "learning_rate": 0.0001622610072455194, + "loss": 3.0347838401794434, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.030014054849743843, + "learning_rate": 0.00016202630902426355, + "loss": 3.0315332412719727, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.02932834066450596, + "learning_rate": 0.00016179172438296423, + "loss": 3.084340810775757, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.029181934893131256, + "learning_rate": 0.00016155725351062205, + "loss": 3.032886028289795, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.02966221794486046, + "learning_rate": 0.00016132289659614605, + "loss": 3.032766819000244, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.029713116586208344, + "learning_rate": 0.00016108865382835318, + "loss": 3.028778076171875, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.029243769124150276, + "learning_rate": 0.00016085452539596864, + "loss": 3.0492146015167236, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.02997511625289917, + "learning_rate": 0.0001606205114876256, + "loss": 3.0772786140441895, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.02930292673408985, + "learning_rate": 0.00016038661229186472, + "loss": 3.055750608444214, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.03037521243095398, + "learning_rate": 0.00016015282799713453, + "loss": 3.033710479736328, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.029957240447402, + "learning_rate": 0.0001599191587917905, + "loss": 3.0394721031188965, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.02925589308142662, + "learning_rate": 0.0001596856048640961, + "loss": 3.0878353118896484, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.03003518097102642, + "learning_rate": 0.00015945216640222116, + "loss": 3.048628330230713, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.031545430421829224, + "learning_rate": 0.000159218843594243, + "loss": 3.0775623321533203, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.031310439109802246, + "learning_rate": 0.0001589856366281455, + "loss": 3.01645565032959, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.03074057213962078, + "learning_rate": 0.0001587525456918194, + "loss": 3.0651440620422363, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.028704499825835228, + "learning_rate": 0.0001585195709730619, + "loss": 3.0527148246765137, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.029061976820230484, + "learning_rate": 0.00015828671265957623, + "loss": 3.0854501724243164, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.03034680150449276, + "learning_rate": 0.00015805397093897255, + "loss": 3.056356906890869, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.028715113177895546, + "learning_rate": 0.00015782134599876633, + "loss": 3.0874247550964355, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.030256496742367744, + "learning_rate": 0.00015758883802637953, + "loss": 3.052424430847168, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.02949833683669567, + "learning_rate": 0.00015735644720913923, + "loss": 3.0663650035858154, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.028980428352952003, + "learning_rate": 0.00015712417373427897, + "loss": 3.0700113773345947, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.028526026755571365, + "learning_rate": 0.00015689201778893713, + "loss": 3.0856711864471436, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.029572727158665657, + "learning_rate": 0.00015665997956015743, + "loss": 3.0862035751342773, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.02955584041774273, + "learning_rate": 0.00015642805923488895, + "loss": 3.030802011489868, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.02881692908704281, + "learning_rate": 0.00015619625699998585, + "loss": 3.074392318725586, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.029564999043941498, + "learning_rate": 0.00015596457304220692, + "loss": 3.057894229888916, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.030315004289150238, + "learning_rate": 0.00015573300754821547, + "loss": 3.096233367919922, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.030666006729006767, + "learning_rate": 0.00015550156070458002, + "loss": 3.040799617767334, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.028901133686304092, + "learning_rate": 0.0001552702326977731, + "loss": 3.055051803588867, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.030359387397766113, + "learning_rate": 0.00015503902371417122, + "loss": 3.085972309112549, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.03117106668651104, + "learning_rate": 0.00015480793394005547, + "loss": 3.0786352157592773, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.029238386079669, + "learning_rate": 0.0001545769635616107, + "loss": 3.0375771522521973, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.029324093833565712, + "learning_rate": 0.0001543461127649256, + "loss": 3.1140198707580566, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.02875387854874134, + "learning_rate": 0.00015411538173599231, + "loss": 3.090881109237671, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.029023943468928337, + "learning_rate": 0.00015388477066070674, + "loss": 3.0755701065063477, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.0291463453322649, + "learning_rate": 0.00015365427972486827, + "loss": 3.0723776817321777, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.02795177884399891, + "learning_rate": 0.00015342390911417904, + "loss": 3.058993339538574, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.029499078169465065, + "learning_rate": 0.00015319365901424452, + "loss": 3.0856404304504395, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.029889609664678574, + "learning_rate": 0.0001529635296105732, + "loss": 3.0622942447662354, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.03107610158622265, + "learning_rate": 0.0001527335210885761, + "loss": 3.114957332611084, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.02967257983982563, + "learning_rate": 0.00015250363363356685, + "loss": 3.088871479034424, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.029068559408187866, + "learning_rate": 0.00015227386743076155, + "loss": 3.1103081703186035, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.029334546998143196, + "learning_rate": 0.00015204422266527911, + "loss": 3.0841622352600098, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.029259610921144485, + "learning_rate": 0.00015181469952213968, + "loss": 3.0662848949432373, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.030072513967752457, + "learning_rate": 0.0001515852981862663, + "loss": 3.0900328159332275, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.033321622759103775, + "learning_rate": 0.00015135601884248303, + "loss": 3.0871129035949707, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.271463394165039, + "eval_runtime": 57.9048, + "eval_samples_per_second": 42.173, + "eval_steps_per_second": 1.33, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.03635408729314804, + "learning_rate": 0.0001511268616755166, + "loss": 2.9576284885406494, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.03927835449576378, + "learning_rate": 0.00015089782686999445, + "loss": 2.999624729156494, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.03371119126677513, + "learning_rate": 0.00015066891461044588, + "loss": 2.971052646636963, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.033049341291189194, + "learning_rate": 0.00015044012508130138, + "loss": 2.980811595916748, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.03830729424953461, + "learning_rate": 0.00015021145846689254, + "loss": 2.9985947608947754, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.036573827266693115, + "learning_rate": 0.00014998291495145206, + "loss": 2.9892568588256836, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.031196987256407738, + "learning_rate": 0.00014975449471911294, + "loss": 2.9854254722595215, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.03376495838165283, + "learning_rate": 0.0001495261979539097, + "loss": 2.9840264320373535, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.036272983998060226, + "learning_rate": 0.0001492980248397766, + "loss": 3.0168871879577637, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.03328242897987366, + "learning_rate": 0.00014906997556054865, + "loss": 2.9815053939819336, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.03158426657319069, + "learning_rate": 0.00014884205029996113, + "loss": 2.994525909423828, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.03320910036563873, + "learning_rate": 0.00014861424924164925, + "loss": 3.008638858795166, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.03160254284739494, + "learning_rate": 0.00014838657256914832, + "loss": 2.979243040084839, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.03158991411328316, + "learning_rate": 0.0001481590204658931, + "loss": 2.938634157180786, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.031116485595703125, + "learning_rate": 0.00014793159311521833, + "loss": 2.9957656860351562, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.030757686123251915, + "learning_rate": 0.00014770429070035812, + "loss": 3.0106635093688965, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.02992827817797661, + "learning_rate": 0.00014747711340444601, + "loss": 2.996793746948242, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.0330309197306633, + "learning_rate": 0.00014725006141051434, + "loss": 2.9859161376953125, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.029488038271665573, + "learning_rate": 0.00014702313490149512, + "loss": 3.0081536769866943, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.03095114603638649, + "learning_rate": 0.00014679633406021892, + "loss": 3.006385087966919, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.03194110840559006, + "learning_rate": 0.00014656965906941494, + "loss": 3.0125250816345215, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.03306785598397255, + "learning_rate": 0.00014634311011171112, + "loss": 2.9640283584594727, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.030883541330695152, + "learning_rate": 0.00014611668736963396, + "loss": 3.0223653316497803, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.03224025294184685, + "learning_rate": 0.00014589039102560822, + "loss": 3.0274910926818848, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.03182433918118477, + "learning_rate": 0.00014566422126195656, + "loss": 3.047736406326294, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.03387221321463585, + "learning_rate": 0.00014543817826089992, + "loss": 2.9821343421936035, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.03407924994826317, + "learning_rate": 0.0001452122622045574, + "loss": 3.041207790374756, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.031546324491500854, + "learning_rate": 0.00014498647327494506, + "loss": 3.0298609733581543, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.03354770317673683, + "learning_rate": 0.00014476081165397715, + "loss": 3.0212411880493164, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.031692177057266235, + "learning_rate": 0.00014453527752346517, + "loss": 3.007350444793701, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.03218817710876465, + "learning_rate": 0.00014430987106511795, + "loss": 3.0019829273223877, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.03233771771192551, + "learning_rate": 0.00014408459246054126, + "loss": 3.0400614738464355, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.028998777270317078, + "learning_rate": 0.00014385944189123794, + "loss": 3.022221088409424, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.030422436073422432, + "learning_rate": 0.00014363441953860805, + "loss": 3.0071797370910645, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.030207227915525436, + "learning_rate": 0.00014340952558394778, + "loss": 2.9762964248657227, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.0305719543248415, + "learning_rate": 0.00014318476020845018, + "loss": 3.075176239013672, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.03173545375466347, + "learning_rate": 0.0001429601235932047, + "loss": 3.028982162475586, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.02799576334655285, + "learning_rate": 0.00014273561591919705, + "loss": 3.0242385864257812, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.03140150383114815, + "learning_rate": 0.00014251123736730884, + "loss": 3.012221336364746, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.029416466131806374, + "learning_rate": 0.00014228698811831787, + "loss": 3.0037949085235596, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.028385788202285767, + "learning_rate": 0.0001420628683528978, + "loss": 3.0350992679595947, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.031597770750522614, + "learning_rate": 0.00014183887825161779, + "loss": 3.065139055252075, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.02926146425306797, + "learning_rate": 0.0001416150179949428, + "loss": 3.014230728149414, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.02970231883227825, + "learning_rate": 0.0001413912877632326, + "loss": 3.06636381149292, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.029684338718652725, + "learning_rate": 0.00014116768773674306, + "loss": 3.024610996246338, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.02908361330628395, + "learning_rate": 0.00014094421809562438, + "loss": 3.0093719959259033, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.029891187325119972, + "learning_rate": 0.00014072087901992209, + "loss": 3.0735411643981934, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.028607673943042755, + "learning_rate": 0.00014049767068957645, + "loss": 3.044274091720581, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.029806701466441154, + "learning_rate": 0.00014027459328442242, + "loss": 3.0449328422546387, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.03161340951919556, + "learning_rate": 0.00014005164698418948, + "loss": 3.046813488006592, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.028860243037343025, + "learning_rate": 0.0001398288319685013, + "loss": 3.011828660964966, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.0285879448056221, + "learning_rate": 0.00013960614841687595, + "loss": 3.043640613555908, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.0296401996165514, + "learning_rate": 0.00013938359650872558, + "loss": 3.071589469909668, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.02807026542723179, + "learning_rate": 0.00013916117642335623, + "loss": 3.018320083618164, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.029037052765488625, + "learning_rate": 0.00013893888833996774, + "loss": 3.0471763610839844, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.029226379469037056, + "learning_rate": 0.00013871673243765362, + "loss": 3.037463426589966, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.029322804883122444, + "learning_rate": 0.00013849470889540098, + "loss": 3.0472755432128906, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.028253905475139618, + "learning_rate": 0.00013827281789208996, + "loss": 3.0265414714813232, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.028875533491373062, + "learning_rate": 0.00013805105960649423, + "loss": 3.063941717147827, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.028460513800382614, + "learning_rate": 0.00013782943421728042, + "loss": 3.0512683391571045, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.02951696142554283, + "learning_rate": 0.00013760794190300827, + "loss": 3.0528573989868164, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.030593490228056908, + "learning_rate": 0.00013738658284212973, + "loss": 3.057919979095459, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.028074532747268677, + "learning_rate": 0.00013716535721299016, + "loss": 3.0263280868530273, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.02955678664147854, + "learning_rate": 0.0001369442651938271, + "loss": 3.081322193145752, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.029724644497036934, + "learning_rate": 0.00013672330696277014, + "loss": 3.059583902359009, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.029713226482272148, + "learning_rate": 0.00013650248269784143, + "loss": 3.018678903579712, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.029758386313915253, + "learning_rate": 0.00013628179257695508, + "loss": 3.0379555225372314, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.028971998021006584, + "learning_rate": 0.00013606123677791727, + "loss": 3.041267156600952, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.02968725562095642, + "learning_rate": 0.00013584081547842565, + "loss": 3.0675506591796875, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.029876509681344032, + "learning_rate": 0.00013562052885606957, + "loss": 3.051586151123047, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.02853119559586048, + "learning_rate": 0.00013540037708833043, + "loss": 3.0335168838500977, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.03374241292476654, + "learning_rate": 0.00013518036035258012, + "loss": 3.027909278869629, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.030695006251335144, + "learning_rate": 0.0001349604788260823, + "loss": 2.9685583114624023, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0339728482067585, + "learning_rate": 0.00013474073268599152, + "loss": 2.936551570892334, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.03372679278254509, + "learning_rate": 0.00013452112210935339, + "loss": 2.9631872177124023, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.03172554448246956, + "learning_rate": 0.00013430164727310399, + "loss": 2.969510078430176, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.030516354367136955, + "learning_rate": 0.00013408230835407027, + "loss": 2.930342674255371, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.03261839970946312, + "learning_rate": 0.00013386310552896977, + "loss": 2.9920477867126465, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.030618635937571526, + "learning_rate": 0.00013364403897441016, + "loss": 2.9693596363067627, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.030576011165976524, + "learning_rate": 0.00013342510886688944, + "loss": 2.929804801940918, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.03224785998463631, + "learning_rate": 0.00013320631538279548, + "loss": 2.95280122756958, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.030605586245656013, + "learning_rate": 0.00013298765869840658, + "loss": 2.931358575820923, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.031206639483571053, + "learning_rate": 0.00013276913898989013, + "loss": 2.962998151779175, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.031317099928855896, + "learning_rate": 0.00013255075643330369, + "loss": 2.9706358909606934, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.031614501029253006, + "learning_rate": 0.00013233251120459407, + "loss": 2.9326977729797363, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.031002728268504143, + "learning_rate": 0.0001321144034795976, + "loss": 2.99625301361084, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.03216253221035004, + "learning_rate": 0.00013189643343403977, + "loss": 2.920912027359009, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.033380527049303055, + "learning_rate": 0.00013167860124353476, + "loss": 2.9688234329223633, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.03085758350789547, + "learning_rate": 0.00013146090708358657, + "loss": 2.9811222553253174, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.03258587792515755, + "learning_rate": 0.00013124335112958704, + "loss": 2.932366371154785, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.031639691442251205, + "learning_rate": 0.0001310259335568172, + "loss": 2.988865613937378, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.03144839406013489, + "learning_rate": 0.00013080865454044647, + "loss": 2.9750723838806152, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.030864672735333443, + "learning_rate": 0.00013059151425553263, + "loss": 2.999329090118408, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.030679022893309593, + "learning_rate": 0.00013037451287702183, + "loss": 2.974264144897461, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.03236627206206322, + "learning_rate": 0.0001301576505797479, + "loss": 2.9808404445648193, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.03115275874733925, + "learning_rate": 0.00012994092753843293, + "loss": 3.0216290950775146, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.029868317767977715, + "learning_rate": 0.00012972434392768687, + "loss": 2.9739654064178467, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.03446527197957039, + "learning_rate": 0.00012950789992200714, + "loss": 2.9565515518188477, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.033536117523908615, + "learning_rate": 0.00012929159569577886, + "loss": 2.958047389984131, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.03113192319869995, + "learning_rate": 0.0001290754314232743, + "loss": 2.9662246704101562, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.03089931420981884, + "learning_rate": 0.00012885940727865334, + "loss": 2.9972119331359863, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.030822310596704483, + "learning_rate": 0.00012864352343596247, + "loss": 3.020500421524048, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.032264843583106995, + "learning_rate": 0.0001284277800691355, + "loss": 2.9895882606506348, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.031226349994540215, + "learning_rate": 0.00012821217735199298, + "loss": 2.958217144012451, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.03110436722636223, + "learning_rate": 0.00012799671545824212, + "loss": 2.9884791374206543, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.03358910232782364, + "learning_rate": 0.0001277813945614768, + "loss": 2.9908103942871094, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.03348501771688461, + "learning_rate": 0.00012756621483517682, + "loss": 2.985677480697632, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.029595404863357544, + "learning_rate": 0.00012735117645270905, + "loss": 2.979318380355835, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.031459368765354156, + "learning_rate": 0.00012713627958732567, + "loss": 2.9921798706054688, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.03067564032971859, + "learning_rate": 0.00012692152441216539, + "loss": 2.977778434753418, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.030506785959005356, + "learning_rate": 0.0001267069111002525, + "loss": 2.9841768741607666, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.029978664591908455, + "learning_rate": 0.00012649243982449718, + "loss": 2.995443344116211, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.030900927260518074, + "learning_rate": 0.00012627811075769502, + "loss": 3.0177016258239746, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.029710259288549423, + "learning_rate": 0.00012606392407252687, + "loss": 3.01532244682312, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.031190430745482445, + "learning_rate": 0.00012584987994155943, + "loss": 2.9791834354400635, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.03150483965873718, + "learning_rate": 0.00012563597853724388, + "loss": 2.998800754547119, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.030249621719121933, + "learning_rate": 0.000125422220031917, + "loss": 3.0187838077545166, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.031052686274051666, + "learning_rate": 0.00012520860459779972, + "loss": 2.984062433242798, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.02961024083197117, + "learning_rate": 0.0001249951324069986, + "loss": 2.970818042755127, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.03020382858812809, + "learning_rate": 0.0001247818036315042, + "loss": 2.9913196563720703, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.031914591789245605, + "learning_rate": 0.00012456861844319155, + "loss": 2.993145704269409, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.03061537817120552, + "learning_rate": 0.00012435557701382012, + "loss": 3.0131571292877197, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.029960963875055313, + "learning_rate": 0.0001241426795150336, + "loss": 3.0017619132995605, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.030348235741257668, + "learning_rate": 0.00012392992611835973, + "loss": 3.0220556259155273, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.03030729480087757, + "learning_rate": 0.0001237173169952098, + "loss": 3.0499038696289062, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.030809229239821434, + "learning_rate": 0.00012350485231687954, + "loss": 2.986651659011841, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.029438305646181107, + "learning_rate": 0.0001232925322545476, + "loss": 3.0351314544677734, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.029887216165661812, + "learning_rate": 0.0001230803569792765, + "loss": 3.021660327911377, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.02955588512122631, + "learning_rate": 0.0001228683266620121, + "loss": 2.9990694522857666, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.029586469754576683, + "learning_rate": 0.00012265644147358326, + "loss": 3.0296952724456787, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.030632557347416878, + "learning_rate": 0.00012244470158470226, + "loss": 3.0124430656433105, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.030579863116145134, + "learning_rate": 0.00012223310716596387, + "loss": 3.0381762981414795, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.03282996267080307, + "learning_rate": 0.00012202165838784601, + "loss": 3.0329041481018066, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.031176412478089333, + "learning_rate": 0.00012181035542070913, + "loss": 3.037598133087158, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.028640294447541237, + "learning_rate": 0.00012159919843479617, + "loss": 3.0004701614379883, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.0326533280313015, + "learning_rate": 0.00012138818760023257, + "loss": 3.0328853130340576, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.03041139990091324, + "learning_rate": 0.00012117732308702592, + "loss": 3.0149893760681152, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.030366195365786552, + "learning_rate": 0.00012096660506506605, + "loss": 3.0379695892333984, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.03106893040239811, + "learning_rate": 0.00012075603370412443, + "loss": 3.0284576416015625, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.0288887657225132, + "learning_rate": 0.00012054560917385476, + "loss": 3.028569221496582, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.03154266998171806, + "learning_rate": 0.00012033533164379225, + "loss": 3.0737850666046143, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.0305064395070076, + "learning_rate": 0.0001201252012833537, + "loss": 3.0744194984436035, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.030967431142926216, + "learning_rate": 0.00011991521826183747, + "loss": 3.0170607566833496, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.036945272237062454, + "learning_rate": 0.00011970538274842282, + "loss": 3.0355610847473145, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.329805374145508, + "eval_runtime": 64.1339, + "eval_samples_per_second": 38.077, + "eval_steps_per_second": 1.201, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.03236053138971329, + "learning_rate": 0.00011949569491217073, + "loss": 2.9230728149414062, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.0395631417632103, + "learning_rate": 0.00011928615492202269, + "loss": 2.92578125, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.0361955352127552, + "learning_rate": 0.0001190767629468014, + "loss": 2.937826633453369, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.035918429493904114, + "learning_rate": 0.00011886751915521007, + "loss": 2.9481167793273926, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.03309318795800209, + "learning_rate": 0.00011865842371583278, + "loss": 2.8919506072998047, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0327315591275692, + "learning_rate": 0.00011844947679713396, + "loss": 2.903585433959961, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.033141329884529114, + "learning_rate": 0.00011824067856745812, + "loss": 2.942802667617798, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.033966805785894394, + "learning_rate": 0.00011803202919503064, + "loss": 2.9210848808288574, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.03218021243810654, + "learning_rate": 0.00011782352884795615, + "loss": 2.933588981628418, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.03272828459739685, + "learning_rate": 0.00011761517769421983, + "loss": 2.9202046394348145, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.030764421448111534, + "learning_rate": 0.00011740697590168635, + "loss": 2.916121006011963, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.031897611916065216, + "learning_rate": 0.00011719892363810018, + "loss": 2.957679033279419, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.03199796751141548, + "learning_rate": 0.00011699102107108539, + "loss": 2.9276621341705322, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.030233927071094513, + "learning_rate": 0.00011678326836814507, + "loss": 2.9006078243255615, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.03035062551498413, + "learning_rate": 0.00011657566569666198, + "loss": 2.920341730117798, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.03029404580593109, + "learning_rate": 0.00011636821322389777, + "loss": 2.937875747680664, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.029371997341513634, + "learning_rate": 0.00011616091111699333, + "loss": 2.9832730293273926, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.030163884162902832, + "learning_rate": 0.00011595375954296786, + "loss": 2.94118332862854, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.03072655200958252, + "learning_rate": 0.00011574675866871997, + "loss": 2.9632580280303955, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.02975962497293949, + "learning_rate": 0.00011553990866102651, + "loss": 2.9346470832824707, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.029814405366778374, + "learning_rate": 0.00011533320968654265, + "loss": 2.9451403617858887, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.030135225504636765, + "learning_rate": 0.00011512666191180204, + "loss": 2.944918155670166, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.029395751655101776, + "learning_rate": 0.00011492026550321653, + "loss": 2.925586223602295, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.0306396484375, + "learning_rate": 0.00011471402062707607, + "loss": 2.9487359523773193, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.03110465221107006, + "learning_rate": 0.00011450792744954827, + "loss": 2.963491678237915, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.03155078738927841, + "learning_rate": 0.00011430198613667861, + "loss": 2.9355576038360596, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.030103784054517746, + "learning_rate": 0.00011409619685439064, + "loss": 2.953045606613159, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.030827803537249565, + "learning_rate": 0.00011389055976848477, + "loss": 2.968574047088623, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.0310911163687706, + "learning_rate": 0.00011368507504463914, + "loss": 2.977902412414551, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.030260782688856125, + "learning_rate": 0.00011347974284840914, + "loss": 2.94873309135437, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.030654428526759148, + "learning_rate": 0.00011327456334522725, + "loss": 2.9282753467559814, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.02993479184806347, + "learning_rate": 0.00011306953670040275, + "loss": 2.97076416015625, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.030244970694184303, + "learning_rate": 0.00011286466307912184, + "loss": 2.960441827774048, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.030003050342202187, + "learning_rate": 0.00011265994264644783, + "loss": 2.9672164916992188, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.030057862401008606, + "learning_rate": 0.00011245537556732001, + "loss": 2.961918830871582, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.029379919171333313, + "learning_rate": 0.00011225096200655442, + "loss": 2.9918394088745117, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.03060895763337612, + "learning_rate": 0.00011204670212884342, + "loss": 2.94891619682312, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.030228130519390106, + "learning_rate": 0.00011184259609875557, + "loss": 2.9726617336273193, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.029833678156137466, + "learning_rate": 0.00011163864408073522, + "loss": 2.9494335651397705, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.030842263251543045, + "learning_rate": 0.00011143484623910293, + "loss": 2.9580483436584473, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.029707176610827446, + "learning_rate": 0.00011123120273805496, + "loss": 2.963684320449829, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.030966786667704582, + "learning_rate": 0.0001110277137416632, + "loss": 2.9710958003997803, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.03137032687664032, + "learning_rate": 0.00011082437941387512, + "loss": 3.003589391708374, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.03123190440237522, + "learning_rate": 0.00011062119991851322, + "loss": 2.9742860794067383, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.030608588829636574, + "learning_rate": 0.000110418175419276, + "loss": 2.97404408454895, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.030018644407391548, + "learning_rate": 0.00011021530607973626, + "loss": 2.972231864929199, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.03056846372783184, + "learning_rate": 0.00011001259206334235, + "loss": 2.9871716499328613, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.02888435684144497, + "learning_rate": 0.00010981003353341721, + "loss": 2.9774169921875, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.03173990175127983, + "learning_rate": 0.00010960763065315864, + "loss": 2.941751480102539, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.030047327280044556, + "learning_rate": 0.00010940538358563906, + "loss": 2.990516185760498, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.030729815363883972, + "learning_rate": 0.00010920329249380509, + "loss": 2.9745054244995117, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.030600983649492264, + "learning_rate": 0.00010900135754047799, + "loss": 2.9676642417907715, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.03208782523870468, + "learning_rate": 0.00010879957888835307, + "loss": 3.006941556930542, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.031162556260824203, + "learning_rate": 0.00010859795669999981, + "loss": 2.9740395545959473, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.031870320439338684, + "learning_rate": 0.00010839649113786137, + "loss": 2.9809296131134033, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.03184300288558006, + "learning_rate": 0.00010819518236425514, + "loss": 2.9781875610351562, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.029629364609718323, + "learning_rate": 0.00010799403054137197, + "loss": 2.9912891387939453, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.03322628512978554, + "learning_rate": 0.00010779303583127609, + "loss": 2.9957847595214844, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.0322413295507431, + "learning_rate": 0.00010759219839590535, + "loss": 2.9894957542419434, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.030753307044506073, + "learning_rate": 0.00010739151839707089, + "loss": 3.0320589542388916, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.03447900339961052, + "learning_rate": 0.00010719099599645706, + "loss": 3.0189106464385986, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.032016605138778687, + "learning_rate": 0.0001069906313556208, + "loss": 3.0162312984466553, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.030838409438729286, + "learning_rate": 0.00010679042463599263, + "loss": 3.000263214111328, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.034556787461042404, + "learning_rate": 0.00010659037599887545, + "loss": 2.996680736541748, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.03266173601150513, + "learning_rate": 0.0001063904856054446, + "loss": 2.990510940551758, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.030973197892308235, + "learning_rate": 0.00010619075361674836, + "loss": 2.9767255783081055, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.03442178666591644, + "learning_rate": 0.0001059911801937071, + "loss": 2.9941515922546387, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.03215927258133888, + "learning_rate": 0.00010579176549711365, + "loss": 3.006618022918701, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.030949831008911133, + "learning_rate": 0.00010559250968763265, + "loss": 3.004136085510254, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.03256693854928017, + "learning_rate": 0.00010539341292580086, + "loss": 3.0229580402374268, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.02986278384923935, + "learning_rate": 0.00010519447537202729, + "loss": 2.9580869674682617, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.03597614914178848, + "learning_rate": 0.00010499569718659201, + "loss": 2.995757579803467, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.359405517578125, + "eval_runtime": 68.1721, + "eval_samples_per_second": 35.821, + "eval_steps_per_second": 1.129, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.03432514891028404, + "learning_rate": 0.00010479707852964713, + "loss": 2.922243118286133, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.0345693901181221, + "learning_rate": 0.00010459861956121611, + "loss": 2.9114043712615967, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.0349399633705616, + "learning_rate": 0.00010440032044119383, + "loss": 2.8668057918548584, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.03161955252289772, + "learning_rate": 0.00010420218132934614, + "loss": 2.8996763229370117, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.032124556601047516, + "learning_rate": 0.00010400420238531023, + "loss": 2.8662075996398926, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.03291172534227371, + "learning_rate": 0.00010380638376859416, + "loss": 2.9088220596313477, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.033552397042512894, + "learning_rate": 0.00010360872563857682, + "loss": 2.9076967239379883, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.03218761831521988, + "learning_rate": 0.0001034112281545079, + "loss": 2.8726117610931396, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.03289451450109482, + "learning_rate": 0.00010321389147550725, + "loss": 2.9127345085144043, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.033019524067640305, + "learning_rate": 0.00010301671576056588, + "loss": 2.920897960662842, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.03153533488512039, + "learning_rate": 0.00010281970116854436, + "loss": 2.9124374389648438, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.032293178141117096, + "learning_rate": 0.00010262284785817392, + "loss": 2.877277374267578, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.031969670206308365, + "learning_rate": 0.00010242615598805574, + "loss": 2.921635150909424, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.03023488074541092, + "learning_rate": 0.00010222962571666088, + "loss": 2.9336061477661133, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.03215106576681137, + "learning_rate": 0.00010203325720233032, + "loss": 2.905877113342285, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.030415233224630356, + "learning_rate": 0.00010183705060327433, + "loss": 2.909331798553467, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.031150560826063156, + "learning_rate": 0.00010164100607757346, + "loss": 2.9302995204925537, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.03152180463075638, + "learning_rate": 0.00010144512378317687, + "loss": 2.9171547889709473, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.02996763028204441, + "learning_rate": 0.00010124940387790354, + "loss": 2.916459798812866, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.030544105917215347, + "learning_rate": 0.00010105384651944148, + "loss": 2.898505449295044, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.030963370576500893, + "learning_rate": 0.00010085845186534769, + "loss": 2.948704719543457, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.03191790357232094, + "learning_rate": 0.00010066322007304819, + "loss": 2.9182960987091064, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.02956460975110531, + "learning_rate": 0.00010046815129983757, + "loss": 2.9211082458496094, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.031227655708789825, + "learning_rate": 0.00010027324570287925, + "loss": 2.942269802093506, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.031859587877988815, + "learning_rate": 0.00010007850343920519, + "loss": 2.9169349670410156, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.03045276179909706, + "learning_rate": 9.988392466571572e-05, + "loss": 2.9240684509277344, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.031577471643686295, + "learning_rate": 9.968950953917945e-05, + "loss": 2.9135894775390625, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.031031791120767593, + "learning_rate": 9.949525821623309e-05, + "loss": 2.913355827331543, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.030012110248208046, + "learning_rate": 9.930117085338156e-05, + "loss": 2.8860886096954346, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.029772799462080002, + "learning_rate": 9.910724760699731e-05, + "loss": 2.974332809448242, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.029961293563246727, + "learning_rate": 9.891348863332093e-05, + "loss": 2.943694591522217, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.029537590220570564, + "learning_rate": 9.871989408846051e-05, + "loss": 2.93180513381958, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.030834559351205826, + "learning_rate": 9.852646412839178e-05, + "loss": 2.9442358016967773, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.02998523600399494, + "learning_rate": 9.833319890895756e-05, + "loss": 2.9336910247802734, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.030030962079763412, + "learning_rate": 9.814009858586817e-05, + "loss": 2.93593692779541, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.030250679701566696, + "learning_rate": 9.794716331470131e-05, + "loss": 2.9370553493499756, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.02942339889705181, + "learning_rate": 9.77543932509012e-05, + "loss": 2.9406495094299316, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.030448907986283302, + "learning_rate": 9.756178854977925e-05, + "loss": 2.9552698135375977, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.03078317642211914, + "learning_rate": 9.736934936651362e-05, + "loss": 2.933631420135498, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.029188675805926323, + "learning_rate": 9.717707585614916e-05, + "loss": 2.944002628326416, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.03153637796640396, + "learning_rate": 9.698496817359698e-05, + "loss": 2.9708166122436523, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.030299564823508263, + "learning_rate": 9.679302647363476e-05, + "loss": 2.9570717811584473, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.03169458732008934, + "learning_rate": 9.660125091090675e-05, + "loss": 2.942774772644043, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.031527288258075714, + "learning_rate": 9.64096416399228e-05, + "loss": 2.9711742401123047, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.0314360186457634, + "learning_rate": 9.621819881505918e-05, + "loss": 2.9509239196777344, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.032102152705192566, + "learning_rate": 9.602692259055767e-05, + "loss": 2.9443840980529785, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.03111223131418228, + "learning_rate": 9.583581312052646e-05, + "loss": 2.972594738006592, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.0322272814810276, + "learning_rate": 9.564487055893867e-05, + "loss": 2.9361982345581055, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.033640168607234955, + "learning_rate": 9.545409505963338e-05, + "loss": 2.9726009368896484, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.02979348786175251, + "learning_rate": 9.526348677631499e-05, + "loss": 2.9978041648864746, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.030464721843600273, + "learning_rate": 9.507304586255311e-05, + "loss": 2.949810028076172, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.03170578181743622, + "learning_rate": 9.488277247178267e-05, + "loss": 2.9677181243896484, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.02996704913675785, + "learning_rate": 9.469266675730319e-05, + "loss": 2.974087953567505, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.030277637764811516, + "learning_rate": 9.450272887227983e-05, + "loss": 2.95644211769104, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.030859284102916718, + "learning_rate": 9.431295896974182e-05, + "loss": 2.957258701324463, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.029737796634435654, + "learning_rate": 9.412335720258341e-05, + "loss": 2.9502854347229004, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.029822375625371933, + "learning_rate": 9.393392372356335e-05, + "loss": 2.9625749588012695, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.030455606058239937, + "learning_rate": 9.374465868530477e-05, + "loss": 2.9669153690338135, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.029491793364286423, + "learning_rate": 9.355556224029515e-05, + "loss": 2.952991247177124, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.030326122418045998, + "learning_rate": 9.336663454088593e-05, + "loss": 2.978355646133423, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.03011198714375496, + "learning_rate": 9.317787573929282e-05, + "loss": 2.976996421813965, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.029053322970867157, + "learning_rate": 9.298928598759541e-05, + "loss": 2.9754371643066406, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.02925761416554451, + "learning_rate": 9.280086543773698e-05, + "loss": 2.94271183013916, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.029678862541913986, + "learning_rate": 9.26126142415247e-05, + "loss": 2.9386143684387207, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.031623054295778275, + "learning_rate": 9.2424532550629e-05, + "loss": 2.9660329818725586, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.030133802443742752, + "learning_rate": 9.223662051658408e-05, + "loss": 2.9580249786376953, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.029482144862413406, + "learning_rate": 9.204887829078709e-05, + "loss": 2.944474697113037, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.030708568170666695, + "learning_rate": 9.186130602449861e-05, + "loss": 2.967862367630005, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.028793595731258392, + "learning_rate": 9.167390386884224e-05, + "loss": 2.9581642150878906, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.03083081543445587, + "learning_rate": 9.148667197480455e-05, + "loss": 2.957261323928833, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.030180038884282112, + "learning_rate": 9.129961049323494e-05, + "loss": 2.9579484462738037, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.034881673753261566, + "learning_rate": 9.111271957484519e-05, + "loss": 2.9007070064544678, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.38953161239624, + "eval_runtime": 71.4964, + "eval_samples_per_second": 34.156, + "eval_steps_per_second": 1.077, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.0335528701543808, + "learning_rate": 9.092599937021031e-05, + "loss": 2.8542537689208984, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.03163984417915344, + "learning_rate": 9.073945002976715e-05, + "loss": 2.864593505859375, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.034005917608737946, + "learning_rate": 9.055307170381522e-05, + "loss": 2.855072498321533, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.03235425800085068, + "learning_rate": 9.036686454251613e-05, + "loss": 2.8504371643066406, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.03171464428305626, + "learning_rate": 9.018082869589369e-05, + "loss": 2.8924291133880615, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.0333196297287941, + "learning_rate": 8.999496431383368e-05, + "loss": 2.8764119148254395, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.03050111047923565, + "learning_rate": 8.980927154608341e-05, + "loss": 2.874119281768799, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.032114267349243164, + "learning_rate": 8.962375054225255e-05, + "loss": 2.9059693813323975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.03141588345170021, + "learning_rate": 8.943840145181177e-05, + "loss": 2.879429340362549, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.03201865404844284, + "learning_rate": 8.925322442409361e-05, + "loss": 2.898519277572632, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.03152940049767494, + "learning_rate": 8.906821960829163e-05, + "loss": 2.866131544113159, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.032520636916160583, + "learning_rate": 8.888338715346113e-05, + "loss": 2.888209819793701, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.03244909271597862, + "learning_rate": 8.869872720851831e-05, + "loss": 2.888913631439209, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.03160687908530235, + "learning_rate": 8.851423992224012e-05, + "loss": 2.905120372772217, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.031518299132585526, + "learning_rate": 8.832992544326473e-05, + "loss": 2.8985815048217773, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.0304370429366827, + "learning_rate": 8.814578392009104e-05, + "loss": 2.8531291484832764, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.03239843621850014, + "learning_rate": 8.796181550107857e-05, + "loss": 2.876098394393921, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.03194732964038849, + "learning_rate": 8.777802033444712e-05, + "loss": 2.908271074295044, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.03145617991685867, + "learning_rate": 8.75943985682774e-05, + "loss": 2.874128580093384, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.031182898208498955, + "learning_rate": 8.741095035051017e-05, + "loss": 2.894904613494873, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.0323818176984787, + "learning_rate": 8.722767582894613e-05, + "loss": 2.9005861282348633, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.030969973653554916, + "learning_rate": 8.704457515124636e-05, + "loss": 2.9236338138580322, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.03238993510603905, + "learning_rate": 8.686164846493176e-05, + "loss": 2.8811020851135254, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.030876649543642998, + "learning_rate": 8.667889591738317e-05, + "loss": 2.8802058696746826, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.032251305878162384, + "learning_rate": 8.649631765584083e-05, + "loss": 2.882112503051758, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.03334304690361023, + "learning_rate": 8.631391382740482e-05, + "loss": 2.9290661811828613, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.03299225866794586, + "learning_rate": 8.613168457903459e-05, + "loss": 2.908268928527832, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.03335456922650337, + "learning_rate": 8.594963005754901e-05, + "loss": 2.9156572818756104, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.030786098912358284, + "learning_rate": 8.57677504096261e-05, + "loss": 2.8981709480285645, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.03379470854997635, + "learning_rate": 8.558604578180301e-05, + "loss": 2.9483864307403564, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.03020016849040985, + "learning_rate": 8.540451632047593e-05, + "loss": 2.908958911895752, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.032774876803159714, + "learning_rate": 8.522316217189972e-05, + "loss": 2.899505138397217, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.030589981004595757, + "learning_rate": 8.504198348218821e-05, + "loss": 2.9257779121398926, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.03366618603467941, + "learning_rate": 8.486098039731384e-05, + "loss": 2.8749284744262695, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.03038695827126503, + "learning_rate": 8.46801530631075e-05, + "loss": 2.9127252101898193, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.03400666266679764, + "learning_rate": 8.449950162525859e-05, + "loss": 2.922375202178955, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.031702589243650436, + "learning_rate": 8.431902622931443e-05, + "loss": 2.8886446952819824, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.03223840892314911, + "learning_rate": 8.413872702068119e-05, + "loss": 2.906278610229492, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.0327419750392437, + "learning_rate": 8.395860414462238e-05, + "loss": 2.9436042308807373, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.030698910355567932, + "learning_rate": 8.377865774625985e-05, + "loss": 2.897615909576416, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.03250030800700188, + "learning_rate": 8.359888797057316e-05, + "loss": 2.928028106689453, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.030623450875282288, + "learning_rate": 8.341929496239958e-05, + "loss": 2.9268975257873535, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.030981825664639473, + "learning_rate": 8.323987886643402e-05, + "loss": 2.938185691833496, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.03139727562665939, + "learning_rate": 8.306063982722855e-05, + "loss": 2.9397618770599365, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.029906271025538445, + "learning_rate": 8.288157798919318e-05, + "loss": 2.921111583709717, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.03110392577946186, + "learning_rate": 8.27026934965945e-05, + "loss": 2.899641513824463, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.03060143068432808, + "learning_rate": 8.25239864935566e-05, + "loss": 2.922635078430176, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.03061371110379696, + "learning_rate": 8.23454571240605e-05, + "loss": 2.925189733505249, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.03080526366829872, + "learning_rate": 8.216710553194416e-05, + "loss": 2.90329647064209, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.030412547290325165, + "learning_rate": 8.198893186090222e-05, + "loss": 2.9444844722747803, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.03024943359196186, + "learning_rate": 8.181093625448585e-05, + "loss": 2.9165825843811035, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.030653616413474083, + "learning_rate": 8.163311885610299e-05, + "loss": 2.9171371459960938, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.03027702495455742, + "learning_rate": 8.145547980901791e-05, + "loss": 2.9399702548980713, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.030466679483652115, + "learning_rate": 8.127801925635126e-05, + "loss": 2.888066291809082, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.02997267246246338, + "learning_rate": 8.110073734107954e-05, + "loss": 2.906545877456665, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.03026283159852028, + "learning_rate": 8.092363420603584e-05, + "loss": 2.9359335899353027, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.0293166171759367, + "learning_rate": 8.074670999390896e-05, + "loss": 2.9118804931640625, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.031117571517825127, + "learning_rate": 8.056996484724342e-05, + "loss": 2.968198299407959, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.03041299246251583, + "learning_rate": 8.039339890843958e-05, + "loss": 2.9333484172821045, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.03037327155470848, + "learning_rate": 8.02170123197535e-05, + "loss": 2.895859956741333, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.03161945194005966, + "learning_rate": 8.004080522329674e-05, + "loss": 2.9491944313049316, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.031058384105563164, + "learning_rate": 7.986477776103601e-05, + "loss": 2.926038980484009, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.030037537217140198, + "learning_rate": 7.968893007479343e-05, + "loss": 2.941394805908203, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.030748063698410988, + "learning_rate": 7.951326230624658e-05, + "loss": 2.998904228210449, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.030324939638376236, + "learning_rate": 7.933777459692756e-05, + "loss": 2.9226560592651367, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.030745431780815125, + "learning_rate": 7.916246708822373e-05, + "loss": 2.977520227432251, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.02952338382601738, + "learning_rate": 7.898733992137715e-05, + "loss": 2.9365828037261963, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.030132673680782318, + "learning_rate": 7.881239323748475e-05, + "loss": 2.9574036598205566, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.030379803851246834, + "learning_rate": 7.863762717749771e-05, + "loss": 2.9329113960266113, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.029670827090740204, + "learning_rate": 7.846304188222189e-05, + "loss": 2.951117515563965, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.029823293909430504, + "learning_rate": 7.828863749231777e-05, + "loss": 2.9586987495422363, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.03546741604804993, + "learning_rate": 7.811441414829958e-05, + "loss": 2.9378662109375, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.417913436889648, + "eval_runtime": 69.6157, + "eval_samples_per_second": 35.078, + "eval_steps_per_second": 1.106, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.033859945833683014, + "learning_rate": 7.794037199053598e-05, + "loss": 2.8493661880493164, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.032931916415691376, + "learning_rate": 7.776651115924959e-05, + "loss": 2.844602108001709, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.03232948109507561, + "learning_rate": 7.759283179451704e-05, + "loss": 2.8463873863220215, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.032815027981996536, + "learning_rate": 7.741933403626848e-05, + "loss": 2.846658706665039, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.03269147500395775, + "learning_rate": 7.724601802428805e-05, + "loss": 2.8286149501800537, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.03337012976408005, + "learning_rate": 7.707288389821329e-05, + "loss": 2.9202351570129395, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.031733594834804535, + "learning_rate": 7.689993179753519e-05, + "loss": 2.8601999282836914, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.032877493649721146, + "learning_rate": 7.672716186159829e-05, + "loss": 2.8566389083862305, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.033852867782115936, + "learning_rate": 7.655457422959993e-05, + "loss": 2.859797477722168, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.03143248334527016, + "learning_rate": 7.638216904059122e-05, + "loss": 2.8411474227905273, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.033198583871126175, + "learning_rate": 7.620994643347559e-05, + "loss": 2.888702392578125, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.029946891590952873, + "learning_rate": 7.60379065470098e-05, + "loss": 2.8332901000976562, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.031660862267017365, + "learning_rate": 7.586604951980326e-05, + "loss": 2.8255481719970703, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.030189910903573036, + "learning_rate": 7.569437549031813e-05, + "loss": 2.8664190769195557, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.03273668885231018, + "learning_rate": 7.55228845968691e-05, + "loss": 2.886531114578247, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.030584124848246574, + "learning_rate": 7.535157697762313e-05, + "loss": 2.872990131378174, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.032543931156396866, + "learning_rate": 7.518045277059979e-05, + "loss": 2.891528606414795, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.031016048043966293, + "learning_rate": 7.500951211367068e-05, + "loss": 2.8759872913360596, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.03190434351563454, + "learning_rate": 7.483875514455979e-05, + "loss": 2.8757526874542236, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.030996622517704964, + "learning_rate": 7.466818200084264e-05, + "loss": 2.862654447555542, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.03289523348212242, + "learning_rate": 7.449779281994712e-05, + "loss": 2.9064857959747314, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.03322640061378479, + "learning_rate": 7.43275877391528e-05, + "loss": 2.8921048641204834, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.0303682629019022, + "learning_rate": 7.415756689559061e-05, + "loss": 2.886458396911621, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.033412594348192215, + "learning_rate": 7.398773042624341e-05, + "loss": 2.8638253211975098, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.030457086861133575, + "learning_rate": 7.381807846794532e-05, + "loss": 2.835578441619873, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.03225106745958328, + "learning_rate": 7.3648611157382e-05, + "loss": 2.8652632236480713, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.03178466483950615, + "learning_rate": 7.34793286310899e-05, + "loss": 2.8796353340148926, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.03075973503291607, + "learning_rate": 7.331023102545716e-05, + "loss": 2.895150661468506, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.02984696812927723, + "learning_rate": 7.314131847672268e-05, + "loss": 2.882878303527832, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.030945835635066032, + "learning_rate": 7.297259112097608e-05, + "loss": 2.898838520050049, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.030904723331332207, + "learning_rate": 7.280404909415801e-05, + "loss": 2.885678291320801, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.030709490180015564, + "learning_rate": 7.263569253205973e-05, + "loss": 2.8645198345184326, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.030528701841831207, + "learning_rate": 7.24675215703232e-05, + "loss": 2.8877408504486084, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.030471473932266235, + "learning_rate": 7.229953634444055e-05, + "loss": 2.9136905670166016, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.030309896916151047, + "learning_rate": 7.213173698975444e-05, + "loss": 2.9238810539245605, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.030605493113398552, + "learning_rate": 7.1964123641458e-05, + "loss": 2.8624460697174072, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.029686832800507545, + "learning_rate": 7.179669643459405e-05, + "loss": 2.9085543155670166, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.030039243400096893, + "learning_rate": 7.162945550405575e-05, + "loss": 2.8624000549316406, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.02953776717185974, + "learning_rate": 7.146240098458613e-05, + "loss": 2.8893938064575195, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.029651308432221413, + "learning_rate": 7.129553301077801e-05, + "loss": 2.914975166320801, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.030902806669473648, + "learning_rate": 7.112885171707378e-05, + "loss": 2.887071132659912, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.02948773466050625, + "learning_rate": 7.096235723776562e-05, + "loss": 2.8893041610717773, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.03049607202410698, + "learning_rate": 7.079604970699514e-05, + "loss": 2.902736186981201, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.030077926814556122, + "learning_rate": 7.062992925875332e-05, + "loss": 2.8936967849731445, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.0303986556828022, + "learning_rate": 7.046399602688045e-05, + "loss": 2.9223172664642334, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.030650658532977104, + "learning_rate": 7.029825014506572e-05, + "loss": 2.9140920639038086, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.029562659561634064, + "learning_rate": 7.013269174684795e-05, + "loss": 2.918179988861084, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.03063463792204857, + "learning_rate": 6.996732096561422e-05, + "loss": 2.9222826957702637, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.030284354463219643, + "learning_rate": 6.980213793460092e-05, + "loss": 2.8685007095336914, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.029791876673698425, + "learning_rate": 6.963714278689304e-05, + "loss": 2.914102554321289, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.030152756720781326, + "learning_rate": 6.947233565542417e-05, + "loss": 2.8890669345855713, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.02957645244896412, + "learning_rate": 6.930771667297653e-05, + "loss": 2.8851962089538574, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.02963421493768692, + "learning_rate": 6.914328597218042e-05, + "loss": 2.9012060165405273, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.029279837384819984, + "learning_rate": 6.8979043685515e-05, + "loss": 2.949984073638916, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.0302849430590868, + "learning_rate": 6.881498994530708e-05, + "loss": 2.9049415588378906, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.029432371258735657, + "learning_rate": 6.865112488373186e-05, + "loss": 2.8919944763183594, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.030542058870196342, + "learning_rate": 6.84874486328125e-05, + "loss": 2.9168801307678223, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.02911970019340515, + "learning_rate": 6.832396132441993e-05, + "loss": 2.9342129230499268, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.030335791409015656, + "learning_rate": 6.816066309027309e-05, + "loss": 2.918611764907837, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.03031272068619728, + "learning_rate": 6.799755406193815e-05, + "loss": 2.8769750595092773, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.030200788751244545, + "learning_rate": 6.783463437082929e-05, + "loss": 2.8994460105895996, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.03032573312520981, + "learning_rate": 6.767190414820792e-05, + "loss": 2.920720100402832, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.03180326521396637, + "learning_rate": 6.750936352518284e-05, + "loss": 2.891145706176758, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.029451677575707436, + "learning_rate": 6.734701263271011e-05, + "loss": 2.9057083129882812, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.030261319130659103, + "learning_rate": 6.718485160159289e-05, + "loss": 2.9132461547851562, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.031257566064596176, + "learning_rate": 6.702288056248145e-05, + "loss": 2.9089136123657227, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.03043205663561821, + "learning_rate": 6.686109964587285e-05, + "loss": 2.912537097930908, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.030708536505699158, + "learning_rate": 6.669950898211106e-05, + "loss": 2.8841137886047363, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.03044111281633377, + "learning_rate": 6.653810870138676e-05, + "loss": 2.921257257461548, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.029576802626252174, + "learning_rate": 6.637689893373729e-05, + "loss": 2.908468723297119, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.03031427040696144, + "learning_rate": 6.621587980904651e-05, + "loss": 2.936445713043213, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.036312974989414215, + "learning_rate": 6.605505145704437e-05, + "loss": 2.8997855186462402, + "step": 2880 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.2278453445328896e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-2880/training_args.bin b/runs/l2r90-baseline/checkpoint-2880/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-2880/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-3240/chat_template.jinja b/runs/l2r90-baseline/checkpoint-3240/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-3240/config.json b/runs/l2r90-baseline/checkpoint-3240/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-3240/generation_config.json b/runs/l2r90-baseline/checkpoint-3240/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-3240/model.safetensors b/runs/l2r90-baseline/checkpoint-3240/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1add19f2f1040d4b57eabaa78be392881ab265c2 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2d5eaf42372ea6e71f162be49d838261a729973ef8e95f101e7c686675e174b +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-3240/optimizer.pt b/runs/l2r90-baseline/checkpoint-3240/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..db4f028221c1446eb4685ff92d34bee00f323160 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23c38996c8869706cb8694119a4e7734e44af55bcabd4ae2508ccd27b24f1d20 +size 1166827898 diff --git a/runs/l2r90-baseline/checkpoint-3240/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-3240/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..03a3cf3ac5f2494e2e8da8d0950115c2ccc72e58 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87ef313648fca4db6856222ba9d5b377ce9c78e1b0b773915470c174d8c1c2c1 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-3240/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-3240/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..1d17deed2fc60410da20d54a2fbecb4d346413b2 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:667ef72d8ecb4d5779e98d21e507df93f664be0f5a602938961f1e875275982c +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-3240/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-3240/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..bcc6f0655376c0e1e22412736914a5726b35023a --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79a799e6c6d9bf9f5dc43e611788e95f712fea0df4ce951955e0ef5ae1737ed7 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-3240/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-3240/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..76a15cd64874a20e370ab87f745baa61cd9ca761 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:063fb8860cd08a9d7dffc329c1b857f2c84f027b83d2d31f9589ce8e5d92bba5 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-3240/scheduler.pt b/runs/l2r90-baseline/checkpoint-3240/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..84479f3138d6041f59e136173e080a0ddbdf81b1 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1045dafc67457b916f466f6a7399234fef3fce01fc96958c0e146dea7784e901 +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-3240/tokenizer.json b/runs/l2r90-baseline/checkpoint-3240/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-3240/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-3240/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-3240/trainer_state.json b/runs/l2r90-baseline/checkpoint-3240/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..17949cd890aca3314385de9746009d0a7febb771 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/trainer_state.json @@ -0,0 +1,23058 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 45.0, + "eval_steps": 500, + "global_step": 3240, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 5.549652099609375, + "eval_runtime": 57.761, + "eval_samples_per_second": 42.278, + "eval_steps_per_second": 1.333, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.03958690166473389, + "learning_rate": 0.0005919487226162231, + "loss": 5.459286689758301, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.03299890458583832, + "learning_rate": 0.0005918869546699877, + "loss": 5.45464563369751, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.028040863573551178, + "learning_rate": 0.0005918249539727509, + "loss": 5.444796562194824, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.028365720063447952, + "learning_rate": 0.0005917627205744655, + "loss": 5.447852611541748, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.02822224237024784, + "learning_rate": 0.0005917002545252719, + "loss": 5.453883171081543, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.036075398325920105, + "learning_rate": 0.0005916375558754977, + "loss": 5.462671279907227, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04683876410126686, + "learning_rate": 0.0005915746246756581, + "loss": 5.443917274475098, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.05457588657736778, + "learning_rate": 0.0005915114609764558, + "loss": 5.444374084472656, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.04250836744904518, + "learning_rate": 0.0005914480648287804, + "loss": 5.435029983520508, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.029477844014763832, + "learning_rate": 0.0005913844362837093, + "loss": 5.421862602233887, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.03673284500837326, + "learning_rate": 0.0005913205753925066, + "loss": 5.430391311645508, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.030489861965179443, + "learning_rate": 0.0005912564822066241, + "loss": 5.389314651489258, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.04037046805024147, + "learning_rate": 0.0005911921567777005, + "loss": 5.425060272216797, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.0445217490196228, + "learning_rate": 0.0005911275991575613, + "loss": 5.437565326690674, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.03885014355182648, + "learning_rate": 0.0005910628093982198, + "loss": 5.410930633544922, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.0334964245557785, + "learning_rate": 0.0005909977875518759, + "loss": 5.412092208862305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.03499286621809006, + "learning_rate": 0.0005909325336709164, + "loss": 5.382216453552246, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.04171644523739815, + "learning_rate": 0.0005908670478079152, + "loss": 5.387385368347168, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.057558078318834305, + "learning_rate": 0.000590801330015633, + "loss": 5.437925815582275, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.053846780210733414, + "learning_rate": 0.0005907353803470177, + "loss": 5.41644811630249, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.027259036898612976, + "learning_rate": 0.0005906691988552034, + "loss": 5.409185409545898, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.03394043818116188, + "learning_rate": 0.0005906027855935115, + "loss": 5.377973556518555, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.03731574863195419, + "learning_rate": 0.0005905361406154501, + "loss": 5.382505893707275, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.04705141484737396, + "learning_rate": 0.0005904692639747137, + "loss": 5.413661003112793, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.044811323285102844, + "learning_rate": 0.0005904021557251837, + "loss": 5.417835235595703, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.04000752419233322, + "learning_rate": 0.0005903348159209277, + "loss": 5.406160354614258, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.038364164531230927, + "learning_rate": 0.0005902672446162007, + "loss": 5.357950687408447, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.030500127002596855, + "learning_rate": 0.0005901994418654432, + "loss": 5.344951629638672, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.030644729733467102, + "learning_rate": 0.0005901314077232829, + "loss": 5.3599348068237305, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.037976790219545364, + "learning_rate": 0.0005900631422445335, + "loss": 5.354084491729736, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.04106011986732483, + "learning_rate": 0.0005899946454841955, + "loss": 5.363678455352783, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.04334832355380058, + "learning_rate": 0.0005899259174974553, + "loss": 5.350959300994873, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.049939028918743134, + "learning_rate": 0.000589856958339686, + "loss": 5.3317155838012695, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.05300058051943779, + "learning_rate": 0.0005897877680664465, + "loss": 5.353314399719238, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.05943184345960617, + "learning_rate": 0.0005897183467334822, + "loss": 5.3683929443359375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.05600789934396744, + "learning_rate": 0.0005896486943967249, + "loss": 5.345516204833984, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.05392718315124512, + "learning_rate": 0.0005895788111122919, + "loss": 5.291389465332031, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.037913475185632706, + "learning_rate": 0.0005895086969364871, + "loss": 5.353423118591309, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.03362123295664787, + "learning_rate": 0.0005894383519258001, + "loss": 5.379500389099121, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032558735460042953, + "learning_rate": 0.0005893677761369066, + "loss": 5.34043025970459, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.04074402526021004, + "learning_rate": 0.0005892969696266683, + "loss": 5.303869247436523, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.04270484298467636, + "learning_rate": 0.0005892259324521328, + "loss": 5.27297830581665, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.04604959115386009, + "learning_rate": 0.0005891546646705334, + "loss": 5.313268184661865, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.04462246224284172, + "learning_rate": 0.0005890831663392891, + "loss": 5.31167459487915, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.04966094344854355, + "learning_rate": 0.0005890114375160052, + "loss": 5.316695690155029, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.04662073776125908, + "learning_rate": 0.0005889394782584718, + "loss": 5.316898822784424, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.04380315542221069, + "learning_rate": 0.0005888672886246656, + "loss": 5.285087585449219, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03239579498767853, + "learning_rate": 0.0005887948686727483, + "loss": 5.304283142089844, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.029690183699131012, + "learning_rate": 0.0005887222184610675, + "loss": 5.305715084075928, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.03444967418909073, + "learning_rate": 0.0005886493380481559, + "loss": 5.2707672119140625, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.03302565962076187, + "learning_rate": 0.0005885762274927322, + "loss": 5.277251243591309, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.028477763757109642, + "learning_rate": 0.0005885028868537, + "loss": 5.295318603515625, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.022988801822066307, + "learning_rate": 0.0005884293161901487, + "loss": 5.249057292938232, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.026115216314792633, + "learning_rate": 0.000588355515561353, + "loss": 5.270636081695557, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.02576332725584507, + "learning_rate": 0.0005882814850267723, + "loss": 5.263096332550049, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.023978352546691895, + "learning_rate": 0.0005882072246460521, + "loss": 5.276605606079102, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.020695513114333153, + "learning_rate": 0.0005881327344790223, + "loss": 5.2141432762146, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01952219195663929, + "learning_rate": 0.0005880580145856986, + "loss": 5.214289665222168, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.022633206099271774, + "learning_rate": 0.0005879830650262813, + "loss": 5.236690998077393, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.023082347586750984, + "learning_rate": 0.0005879078858611557, + "loss": 5.2092156410217285, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.019620463252067566, + "learning_rate": 0.0005878324771508927, + "loss": 5.264451503753662, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.021504493430256844, + "learning_rate": 0.0005877568389562475, + "loss": 5.23444938659668, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.025672364979982376, + "learning_rate": 0.0005876809713381606, + "loss": 5.205289363861084, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.034436631947755814, + "learning_rate": 0.0005876048743577569, + "loss": 5.2047648429870605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.042389530688524246, + "learning_rate": 0.0005875285480763466, + "loss": 5.192008018493652, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.03310716897249222, + "learning_rate": 0.0005874519925554244, + "loss": 5.162734031677246, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.03029288910329342, + "learning_rate": 0.0005873752078566694, + "loss": 5.212151527404785, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04619310051202774, + "learning_rate": 0.0005872981940419459, + "loss": 5.215909004211426, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.05618087574839592, + "learning_rate": 0.0005872209511733027, + "loss": 5.210192680358887, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.03783707693219185, + "learning_rate": 0.0005871434793129726, + "loss": 5.201597690582275, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.03721104562282562, + "learning_rate": 0.0005870657785233735, + "loss": 5.188891410827637, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.033336155116558075, + "learning_rate": 0.0005869878488671075, + "loss": 5.1598615646362305, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.2592363357543945, + "eval_runtime": 57.3343, + "eval_samples_per_second": 42.592, + "eval_steps_per_second": 1.343, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03719741478562355, + "learning_rate": 0.0005869096904069611, + "loss": 5.1646623611450195, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.04888090118765831, + "learning_rate": 0.0005868313032059052, + "loss": 5.171182155609131, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.05073149502277374, + "learning_rate": 0.0005867526873270949, + "loss": 5.183949947357178, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.04187731072306633, + "learning_rate": 0.0005866738428338695, + "loss": 5.176133155822754, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.039445552974939346, + "learning_rate": 0.000586594769789753, + "loss": 5.148100852966309, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.035551708191633224, + "learning_rate": 0.0005865154682584524, + "loss": 5.151512145996094, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.031175289303064346, + "learning_rate": 0.0005864359383038602, + "loss": 5.102468967437744, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.03696789592504501, + "learning_rate": 0.000586356179990052, + "loss": 5.119902610778809, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.04836135357618332, + "learning_rate": 0.0005862761933812875, + "loss": 5.15183162689209, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.05206914246082306, + "learning_rate": 0.0005861959785420106, + "loss": 5.154870986938477, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.046368084847927094, + "learning_rate": 0.000586115535536849, + "loss": 5.130651473999023, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.07146432995796204, + "learning_rate": 0.000586034864430614, + "loss": 5.175804138183594, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.08809789270162582, + "learning_rate": 0.0005859539652883008, + "loss": 5.229716777801514, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.08116360008716583, + "learning_rate": 0.0005858728381750888, + "loss": 5.21910285949707, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.08282570540904999, + "learning_rate": 0.0005857914831563402, + "loss": 5.2336506843566895, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.07403547316789627, + "learning_rate": 0.0005857099002976016, + "loss": 5.251604080200195, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.07024004310369492, + "learning_rate": 0.0005856280896646026, + "loss": 5.267114162445068, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.056330062448978424, + "learning_rate": 0.0005855460513232567, + "loss": 5.201976299285889, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.051763955503702164, + "learning_rate": 0.0005854637853396606, + "loss": 5.1733198165893555, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.039900992065668106, + "learning_rate": 0.0005853812917800945, + "loss": 5.222941875457764, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.04187863692641258, + "learning_rate": 0.0005852985707110221, + "loss": 5.165033340454102, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.042645812034606934, + "learning_rate": 0.0005852156221990901, + "loss": 5.157554626464844, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.03856075555086136, + "learning_rate": 0.0005851324463111289, + "loss": 5.184629440307617, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.029387330636382103, + "learning_rate": 0.0005850490431141516, + "loss": 5.118739128112793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.027584781870245934, + "learning_rate": 0.0005849654126753545, + "loss": 5.100855350494385, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.023655924946069717, + "learning_rate": 0.0005848815550621175, + "loss": 5.154520034790039, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.024163050577044487, + "learning_rate": 0.0005847974703420028, + "loss": 5.105189323425293, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.020369920879602432, + "learning_rate": 0.0005847131585827564, + "loss": 5.113671779632568, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.021030737087130547, + "learning_rate": 0.0005846286198523061, + "loss": 5.0847649574279785, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.01729249767959118, + "learning_rate": 0.0005845438542187638, + "loss": 5.053225517272949, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.018704712390899658, + "learning_rate": 0.0005844588617504236, + "loss": 5.079106330871582, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.02049064077436924, + "learning_rate": 0.0005843736425157621, + "loss": 5.044185161590576, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.020736856386065483, + "learning_rate": 0.000584288196583439, + "loss": 5.041932106018066, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.016543276607990265, + "learning_rate": 0.0005842025240222966, + "loss": 5.032522201538086, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.018414106220006943, + "learning_rate": 0.0005841166249013598, + "loss": 5.037397384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.01611698977649212, + "learning_rate": 0.0005840304992898359, + "loss": 5.038179397583008, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.014310561120510101, + "learning_rate": 0.0005839441472571147, + "loss": 5.0146484375, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.015472214668989182, + "learning_rate": 0.0005838575688727685, + "loss": 5.060323715209961, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.016373401507735252, + "learning_rate": 0.0005837707642065518, + "loss": 5.032100677490234, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.018694665282964706, + "learning_rate": 0.000583683733328402, + "loss": 5.036993026733398, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.019977087154984474, + "learning_rate": 0.0005835964763084378, + "loss": 4.991037368774414, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.023741086944937706, + "learning_rate": 0.000583508993216961, + "loss": 5.043832778930664, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.025061562657356262, + "learning_rate": 0.0005834212841244548, + "loss": 5.016516208648682, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.020848384127020836, + "learning_rate": 0.0005833333491015851, + "loss": 4.986289978027344, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.015086354687809944, + "learning_rate": 0.0005832451882191995, + "loss": 4.9948625564575195, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.015096964314579964, + "learning_rate": 0.0005831568015483274, + "loss": 4.996342658996582, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016932200640439987, + "learning_rate": 0.0005830681891601807, + "loss": 4.998758316040039, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.019936930388212204, + "learning_rate": 0.0005829793511261525, + "loss": 4.959854602813721, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.027658682316541672, + "learning_rate": 0.0005828902875178183, + "loss": 4.968629837036133, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.03299275413155556, + "learning_rate": 0.0005828009984069347, + "loss": 4.994814872741699, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.033717650920152664, + "learning_rate": 0.0005827114838654406, + "loss": 4.976295471191406, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.02969924360513687, + "learning_rate": 0.0005826217439654562, + "loss": 4.969155788421631, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.02846827358007431, + "learning_rate": 0.0005825317787792831, + "loss": 4.989721298217773, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.03670600801706314, + "learning_rate": 0.0005824415883794049, + "loss": 4.988689422607422, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.045570675283670425, + "learning_rate": 0.0005823511728384863, + "loss": 4.974407196044922, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.04564111679792404, + "learning_rate": 0.0005822605322293733, + "loss": 4.966331481933594, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.04173210263252258, + "learning_rate": 0.0005821696666250937, + "loss": 4.925197601318359, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.04228242486715317, + "learning_rate": 0.0005820785760988559, + "loss": 4.965020656585693, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.05144782364368439, + "learning_rate": 0.0005819872607240503, + "loss": 5.0022149085998535, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.049400657415390015, + "learning_rate": 0.0005818957205742478, + "loss": 4.9635114669799805, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.043692708015441895, + "learning_rate": 0.0005818039557232005, + "loss": 4.977723121643066, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.04891742020845413, + "learning_rate": 0.0005817119662448421, + "loss": 4.996312141418457, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.035564981400966644, + "learning_rate": 0.0005816197522132866, + "loss": 4.940932273864746, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04679827764630318, + "learning_rate": 0.0005815273137028292, + "loss": 4.983131408691406, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.05489890277385712, + "learning_rate": 0.0005814346507879459, + "loss": 4.977827548980713, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04640809819102287, + "learning_rate": 0.0005813417635432937, + "loss": 4.954761505126953, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04439452290534973, + "learning_rate": 0.00058124865204371, + "loss": 4.914865970611572, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.04587096720933914, + "learning_rate": 0.0005811553163642131, + "loss": 4.949592590332031, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.044119734317064285, + "learning_rate": 0.0005810617565800019, + "loss": 4.956068992614746, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.043133001774549484, + "learning_rate": 0.0005809679727664559, + "loss": 4.993522644042969, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04315278306603432, + "learning_rate": 0.0005808739649991348, + "loss": 4.894291877746582, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04819472134113312, + "learning_rate": 0.0005807797333537792, + "loss": 4.958306789398193, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.023706436157227, + "eval_runtime": 57.4082, + "eval_samples_per_second": 42.537, + "eval_steps_per_second": 1.341, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.04403403028845787, + "learning_rate": 0.0005806852779063098, + "loss": 4.884971618652344, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.041993748396635056, + "learning_rate": 0.0005805905987328275, + "loss": 4.916353225708008, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.042302004992961884, + "learning_rate": 0.0005804956959096138, + "loss": 4.9012556076049805, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.044500596821308136, + "learning_rate": 0.0005804005695131299, + "loss": 4.915240287780762, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.03914736211299896, + "learning_rate": 0.0005803052196200177, + "loss": 4.885915756225586, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.0361703559756279, + "learning_rate": 0.0005802096463070988, + "loss": 4.829287528991699, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03432302549481392, + "learning_rate": 0.0005801138496513749, + "loss": 4.926834583282471, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.0317804180085659, + "learning_rate": 0.0005800178297300276, + "loss": 4.8707170486450195, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.035261351615190506, + "learning_rate": 0.0005799215866204183, + "loss": 4.897791862487793, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.03276067599654198, + "learning_rate": 0.0005798251204000886, + "loss": 4.864504814147949, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.034757066518068314, + "learning_rate": 0.0005797284311467594, + "loss": 4.890812873840332, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.033726081252098083, + "learning_rate": 0.0005796315189383316, + "loss": 4.831684112548828, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.03074726276099682, + "learning_rate": 0.0005795343838528855, + "loss": 4.8534111976623535, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02215893566608429, + "learning_rate": 0.0005794370259686811, + "loss": 4.8311238288879395, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.02132033370435238, + "learning_rate": 0.0005793394453641581, + "loss": 4.842617034912109, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020825816318392754, + "learning_rate": 0.0005792416421179351, + "loss": 4.84535026550293, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.019334357231855392, + "learning_rate": 0.0005791436163088106, + "loss": 4.839937686920166, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017372334375977516, + "learning_rate": 0.0005790453680157622, + "loss": 4.8115553855896, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.0190406683832407, + "learning_rate": 0.0005789468973179468, + "loss": 4.832542419433594, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.017372574657201767, + "learning_rate": 0.0005788482042947004, + "loss": 4.793632507324219, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.019171632826328278, + "learning_rate": 0.0005787492890255382, + "loss": 4.776987552642822, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.02109278179705143, + "learning_rate": 0.0005786501515901545, + "loss": 4.8216657638549805, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.026234494522213936, + "learning_rate": 0.0005785507920684225, + "loss": 4.8109049797058105, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.0357641875743866, + "learning_rate": 0.0005784512105403944, + "loss": 4.821109771728516, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.03627947345376015, + "learning_rate": 0.000578351407086301, + "loss": 4.810632228851318, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.032253146171569824, + "learning_rate": 0.0005782513817865527, + "loss": 4.7930521965026855, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.03402210772037506, + "learning_rate": 0.0005781511347217375, + "loss": 4.772500991821289, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.0321584977209568, + "learning_rate": 0.000578050665972623, + "loss": 4.781399726867676, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.030242707580327988, + "learning_rate": 0.0005779499756201549, + "loss": 4.802542686462402, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.02361336164176464, + "learning_rate": 0.0005778490637454576, + "loss": 4.771912574768066, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.028781302273273468, + "learning_rate": 0.000577747930429834, + "loss": 4.796924591064453, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.027042683213949203, + "learning_rate": 0.0005776465757547654, + "loss": 4.745218276977539, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.02715984545648098, + "learning_rate": 0.0005775449998019113, + "loss": 4.792944431304932, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.0314096100628376, + "learning_rate": 0.00057744320265311, + "loss": 4.7649993896484375, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.028749193996191025, + "learning_rate": 0.0005773411843903773, + "loss": 4.7980546951293945, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.02655309997498989, + "learning_rate": 0.0005772389450959075, + "loss": 4.746520519256592, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.027497397735714912, + "learning_rate": 0.000577136484852073, + "loss": 4.7490715980529785, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.0343393050134182, + "learning_rate": 0.000577033803741424, + "loss": 4.74436092376709, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.04444412514567375, + "learning_rate": 0.0005769309018466891, + "loss": 4.755973815917969, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.0437556728720665, + "learning_rate": 0.0005768277792507744, + "loss": 4.784368515014648, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0369921550154686, + "learning_rate": 0.0005767244360367638, + "loss": 4.753410339355469, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.05038141459226608, + "learning_rate": 0.0005766208722879192, + "loss": 4.755491256713867, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.04511422663927078, + "learning_rate": 0.00057651708808768, + "loss": 4.760293006896973, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.048762883991003036, + "learning_rate": 0.0005764130835196632, + "loss": 4.755958080291748, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.06749686598777771, + "learning_rate": 0.0005763088586676634, + "loss": 4.802557468414307, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.06451250612735748, + "learning_rate": 0.0005762044136156528, + "loss": 4.812925338745117, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.05436156690120697, + "learning_rate": 0.0005760997484477809, + "loss": 4.807103157043457, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.05233754590153694, + "learning_rate": 0.0005759948632483742, + "loss": 4.782449722290039, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.05519746616482735, + "learning_rate": 0.0005758897581019371, + "loss": 4.808449745178223, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.0578567236661911, + "learning_rate": 0.000575784433093151, + "loss": 4.7676005363464355, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.0588264986872673, + "learning_rate": 0.0005756788883068743, + "loss": 4.805377960205078, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04225790873169899, + "learning_rate": 0.0005755731238281423, + "loss": 4.748775005340576, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.0376368910074234, + "learning_rate": 0.0005754671397421678, + "loss": 4.759791374206543, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03830771893262863, + "learning_rate": 0.0005753609361343402, + "loss": 4.729974746704102, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.03684520348906517, + "learning_rate": 0.0005752545130902256, + "loss": 4.729854583740234, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02983408235013485, + "learning_rate": 0.0005751478706955674, + "loss": 4.78502082824707, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02854904904961586, + "learning_rate": 0.0005750410090362854, + "loss": 4.717472076416016, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.023819150403141975, + "learning_rate": 0.0005749339281984761, + "loss": 4.717124938964844, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.02477158047258854, + "learning_rate": 0.0005748266282684124, + "loss": 4.71781063079834, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.022002534940838814, + "learning_rate": 0.0005747191093325443, + "loss": 4.7480669021606445, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.02133483812212944, + "learning_rate": 0.0005746113714774976, + "loss": 4.693770408630371, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.022407138720154762, + "learning_rate": 0.0005745034147900747, + "loss": 4.720734119415283, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.020850926637649536, + "learning_rate": 0.0005743952393572544, + "loss": 4.745505332946777, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.020420582965016365, + "learning_rate": 0.0005742868452661918, + "loss": 4.688610076904297, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.019014107063412666, + "learning_rate": 0.0005741782326042181, + "loss": 4.673436164855957, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.017393039539456367, + "learning_rate": 0.0005740694014588403, + "loss": 4.6941118240356445, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.017877329140901566, + "learning_rate": 0.0005739603519177419, + "loss": 4.667474269866943, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.017749635502696037, + "learning_rate": 0.0005738510840687821, + "loss": 4.6901092529296875, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.018746376037597656, + "learning_rate": 0.000573741597999996, + "loss": 4.680166721343994, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016260217875242233, + "learning_rate": 0.0005736318937995947, + "loss": 4.6799821853637695, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015099814161658287, + "learning_rate": 0.0005735219715559646, + "loss": 4.680666923522949, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.016539735719561577, + "learning_rate": 0.0005734118313576683, + "loss": 4.639509201049805, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 4.768754482269287, + "eval_runtime": 58.3352, + "eval_samples_per_second": 41.862, + "eval_steps_per_second": 1.32, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.019733518362045288, + "learning_rate": 0.0005733014732934436, + "loss": 4.635312080383301, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.022571036592125893, + "learning_rate": 0.0005731908974522042, + "loss": 4.6041765213012695, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.022238552570343018, + "learning_rate": 0.0005730801039230389, + "loss": 4.648976802825928, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.02115180902183056, + "learning_rate": 0.0005729690927952123, + "loss": 4.623968124389648, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.021789928898215294, + "learning_rate": 0.0005728578641581637, + "loss": 4.620028018951416, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.02593887411057949, + "learning_rate": 0.0005727464181015081, + "loss": 4.591062545776367, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.0347842313349247, + "learning_rate": 0.0005726347547150357, + "loss": 4.632149696350098, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04119519516825676, + "learning_rate": 0.0005725228740887117, + "loss": 4.6254191398620605, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.04895750433206558, + "learning_rate": 0.0005724107763126761, + "loss": 4.621372222900391, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.05752161890268326, + "learning_rate": 0.0005722984614772442, + "loss": 4.61965274810791, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05115870013833046, + "learning_rate": 0.000572185929672906, + "loss": 4.63119649887085, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.05721089243888855, + "learning_rate": 0.0005720731809903263, + "loss": 4.652647018432617, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.0646839588880539, + "learning_rate": 0.000571960215520345, + "loss": 4.642184257507324, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.0682854950428009, + "learning_rate": 0.0005718470333539757, + "loss": 4.676589012145996, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.057402849197387695, + "learning_rate": 0.0005717336345824077, + "loss": 4.6585283279418945, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.05885373800992966, + "learning_rate": 0.0005716200192970043, + "loss": 4.671728134155273, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.05265705659985542, + "learning_rate": 0.0005715061875893032, + "loss": 4.641249656677246, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.05417842045426369, + "learning_rate": 0.0005713921395510166, + "loss": 4.661381721496582, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.06602369248867035, + "learning_rate": 0.0005712778752740307, + "loss": 4.684137344360352, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.06004182621836662, + "learning_rate": 0.0005711633948504066, + "loss": 4.625332832336426, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.04801511764526367, + "learning_rate": 0.0005710486983723787, + "loss": 4.664791107177734, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.04380105435848236, + "learning_rate": 0.0005709337859323561, + "loss": 4.664966583251953, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.03858727216720581, + "learning_rate": 0.0005708186576229218, + "loss": 4.661867618560791, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03240576758980751, + "learning_rate": 0.0005707033135368323, + "loss": 4.5855326652526855, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.028561925515532494, + "learning_rate": 0.0005705877537670184, + "loss": 4.637331008911133, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.026389246806502342, + "learning_rate": 0.0005704719784065846, + "loss": 4.598666667938232, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.02484819106757641, + "learning_rate": 0.0005703559875488088, + "loss": 4.601729393005371, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.02522977627813816, + "learning_rate": 0.0005702397812871429, + "loss": 4.609833717346191, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02589508146047592, + "learning_rate": 0.0005701233597152121, + "loss": 4.574592590332031, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.023374728858470917, + "learning_rate": 0.0005700067229268154, + "loss": 4.61362361907959, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.02465851791203022, + "learning_rate": 0.0005698898710159245, + "loss": 4.639899253845215, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.02258715033531189, + "learning_rate": 0.0005697728040766852, + "loss": 4.5656538009643555, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.018009789288043976, + "learning_rate": 0.0005696555222034162, + "loss": 4.61449670791626, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02121281623840332, + "learning_rate": 0.0005695380254906094, + "loss": 4.553830623626709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.018078701570630074, + "learning_rate": 0.0005694203140329296, + "loss": 4.534579277038574, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.0169303547590971, + "learning_rate": 0.000569302387925215, + "loss": 4.533995628356934, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.016318142414093018, + "learning_rate": 0.0005691842472624764, + "loss": 4.544438362121582, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.016664575785398483, + "learning_rate": 0.0005690658921398977, + "loss": 4.5913286209106445, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.01641504280269146, + "learning_rate": 0.0005689473226528354, + "loss": 4.520743370056152, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.017012735828757286, + "learning_rate": 0.0005688285388968187, + "loss": 4.517453193664551, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01672312617301941, + "learning_rate": 0.0005687095409675499, + "loss": 4.5235915184021, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.016893865540623665, + "learning_rate": 0.000568590328960903, + "loss": 4.543200492858887, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.017530135810375214, + "learning_rate": 0.0005684709029729253, + "loss": 4.54912805557251, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.018698599189519882, + "learning_rate": 0.0005683512630998361, + "loss": 4.522650718688965, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01984071172773838, + "learning_rate": 0.000568231409438027, + "loss": 4.546173095703125, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.023218289017677307, + "learning_rate": 0.0005681113420840619, + "loss": 4.5305094718933105, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.028968270868062973, + "learning_rate": 0.000567991061134677, + "loss": 4.521440505981445, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.032370273023843765, + "learning_rate": 0.0005678705666867805, + "loss": 4.54168701171875, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.033967748284339905, + "learning_rate": 0.0005677498588374524, + "loss": 4.538717269897461, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.028757929801940918, + "learning_rate": 0.0005676289376839452, + "loss": 4.517228603363037, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.029757825657725334, + "learning_rate": 0.0005675078033236827, + "loss": 4.510310173034668, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.03581003099679947, + "learning_rate": 0.0005673864558542605, + "loss": 4.506901264190674, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.03483985736966133, + "learning_rate": 0.0005672648953734462, + "loss": 4.549284934997559, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.031016027554869652, + "learning_rate": 0.000567143121979179, + "loss": 4.530084609985352, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.029373526573181152, + "learning_rate": 0.0005670211357695693, + "loss": 4.492218971252441, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.031661372631788254, + "learning_rate": 0.0005668989368428994, + "loss": 4.502824783325195, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.027251387014985085, + "learning_rate": 0.0005667765252976227, + "loss": 4.49193811416626, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02263176068663597, + "learning_rate": 0.0005666539012323639, + "loss": 4.512019157409668, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.02355228364467621, + "learning_rate": 0.0005665310647459189, + "loss": 4.508917331695557, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.023297371342778206, + "learning_rate": 0.0005664080159372551, + "loss": 4.497771263122559, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.019364114850759506, + "learning_rate": 0.0005662847549055107, + "loss": 4.51470947265625, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.023899901658296585, + "learning_rate": 0.0005661612817499947, + "loss": 4.48771858215332, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.02330983616411686, + "learning_rate": 0.0005660375965701872, + "loss": 4.478496551513672, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.027252234518527985, + "learning_rate": 0.0005659136994657392, + "loss": 4.497672080993652, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.030886363238096237, + "learning_rate": 0.0005657895905364723, + "loss": 4.49789571762085, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02515559457242489, + "learning_rate": 0.0005656652698823788, + "loss": 4.490771293640137, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020107125863432884, + "learning_rate": 0.0005655407376036217, + "loss": 4.481879234313965, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.022076716646552086, + "learning_rate": 0.0005654159938005345, + "loss": 4.494697570800781, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.025525815784931183, + "learning_rate": 0.0005652910385736207, + "loss": 4.517453193664551, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.03005894646048546, + "learning_rate": 0.0005651658720235548, + "loss": 4.453815460205078, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.030178502202033997, + "learning_rate": 0.0005650404942511812, + "loss": 4.493922233581543, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.029249394312500954, + "learning_rate": 0.0005649149053575141, + "loss": 4.453763008117676, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 4.607339382171631, + "eval_runtime": 57.9216, + "eval_samples_per_second": 42.16, + "eval_steps_per_second": 1.329, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.03362950682640076, + "learning_rate": 0.0005647891054437386, + "loss": 4.430700302124023, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.04223684221506119, + "learning_rate": 0.0005646630946112096, + "loss": 4.486915588378906, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.05192171037197113, + "learning_rate": 0.0005645368729614512, + "loss": 4.467541694641113, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.042525868862867355, + "learning_rate": 0.0005644104405961582, + "loss": 4.45252799987793, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.0406370609998703, + "learning_rate": 0.0005642837976171949, + "loss": 4.414548873901367, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.044476233422756195, + "learning_rate": 0.0005641569441265952, + "loss": 4.453492164611816, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.04687326401472092, + "learning_rate": 0.0005640298802265626, + "loss": 4.477725028991699, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.04011611267924309, + "learning_rate": 0.0005639026060194704, + "loss": 4.423356533050537, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.04971884936094284, + "learning_rate": 0.0005637751216078609, + "loss": 4.4853515625, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.05863234028220177, + "learning_rate": 0.0005636474270944461, + "loss": 4.466001033782959, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.057710036635398865, + "learning_rate": 0.0005635195225821072, + "loss": 4.466723442077637, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.05155074968934059, + "learning_rate": 0.0005633914081738945, + "loss": 4.4823198318481445, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.04535634070634842, + "learning_rate": 0.0005632630839730275, + "loss": 4.471219062805176, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.045147854834795, + "learning_rate": 0.0005631345500828946, + "loss": 4.476888179779053, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.03730619326233864, + "learning_rate": 0.0005630058066070533, + "loss": 4.428932189941406, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.034599728882312775, + "learning_rate": 0.0005628768536492299, + "loss": 4.47627592086792, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.032904356718063354, + "learning_rate": 0.0005627476913133193, + "loss": 4.422671794891357, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.03356337174773216, + "learning_rate": 0.0005626183197033856, + "loss": 4.456406593322754, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.035042595118284225, + "learning_rate": 0.0005624887389236609, + "loss": 4.371820449829102, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.03269032761454582, + "learning_rate": 0.0005623589490785462, + "loss": 4.402605056762695, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.028261031955480576, + "learning_rate": 0.0005622289502726108, + "loss": 4.433136940002441, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03301700949668884, + "learning_rate": 0.0005620987426105925, + "loss": 4.388359546661377, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.031664177775382996, + "learning_rate": 0.000561968326197397, + "loss": 4.41859245300293, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.026281699538230896, + "learning_rate": 0.0005618377011380988, + "loss": 4.406128883361816, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.023097911849617958, + "learning_rate": 0.0005617068675379398, + "loss": 4.389717102050781, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.022349286824464798, + "learning_rate": 0.0005615758255023306, + "loss": 4.408173561096191, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.021958226338028908, + "learning_rate": 0.0005614445751368491, + "loss": 4.434830188751221, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024551713839173317, + "learning_rate": 0.0005613131165472415, + "loss": 4.386944770812988, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.028048967942595482, + "learning_rate": 0.0005611814498394216, + "loss": 4.402279376983643, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02733907848596573, + "learning_rate": 0.0005610495751194708, + "loss": 4.403069496154785, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.020858559757471085, + "learning_rate": 0.0005609174924936384, + "loss": 4.406920909881592, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.021491149440407753, + "learning_rate": 0.0005607852020683409, + "loss": 4.385799407958984, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.021495329216122627, + "learning_rate": 0.0005606527039501621, + "loss": 4.4252495765686035, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.019708355888724327, + "learning_rate": 0.0005605199982458535, + "loss": 4.388864994049072, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.01886274851858616, + "learning_rate": 0.0005603870850623338, + "loss": 4.335962295532227, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.01955341547727585, + "learning_rate": 0.0005602539645066884, + "loss": 4.373560905456543, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.018962012603878975, + "learning_rate": 0.0005601206366861706, + "loss": 4.380022048950195, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.016663335263729095, + "learning_rate": 0.0005599871017081999, + "loss": 4.421034336090088, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01676873490214348, + "learning_rate": 0.0005598533596803631, + "loss": 4.382775783538818, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.014379126951098442, + "learning_rate": 0.0005597194107104137, + "loss": 4.41373872756958, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.015897583216428757, + "learning_rate": 0.000559585254906272, + "loss": 4.377284049987793, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.016513120383024216, + "learning_rate": 0.000559450892376025, + "loss": 4.41200065612793, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.016625918447971344, + "learning_rate": 0.000559316323227926, + "loss": 4.381187438964844, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.016028188169002533, + "learning_rate": 0.000559181547570395, + "loss": 4.353909015655518, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01682914048433304, + "learning_rate": 0.0005590465655120183, + "loss": 4.382326126098633, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.017078889533877373, + "learning_rate": 0.0005589113771615486, + "loss": 4.334949493408203, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.018182938918471336, + "learning_rate": 0.0005587759826279046, + "loss": 4.348130226135254, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01940886303782463, + "learning_rate": 0.0005586403820201713, + "loss": 4.339404106140137, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.020644692704081535, + "learning_rate": 0.0005585045754475995, + "loss": 4.358404159545898, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02433430217206478, + "learning_rate": 0.0005583685630196064, + "loss": 4.3660502433776855, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.027824537828564644, + "learning_rate": 0.0005582323448457744, + "loss": 4.34050178527832, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.02973582223057747, + "learning_rate": 0.0005580959210358523, + "loss": 4.348322868347168, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.030396128073334694, + "learning_rate": 0.0005579592916997542, + "loss": 4.356224060058594, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.029760530218482018, + "learning_rate": 0.0005578224569475599, + "loss": 4.357470989227295, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.02855013869702816, + "learning_rate": 0.0005576854168895147, + "loss": 4.335353374481201, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03376635164022446, + "learning_rate": 0.0005575481716360292, + "loss": 4.331794738769531, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.034431129693984985, + "learning_rate": 0.0005574107212976796, + "loss": 4.336880207061768, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.03524135425686836, + "learning_rate": 0.000557273065985207, + "loss": 4.31537389755249, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.03152058273553848, + "learning_rate": 0.0005571352058095179, + "loss": 4.37662410736084, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02878640592098236, + "learning_rate": 0.0005569971408816838, + "loss": 4.292388916015625, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.029663654044270515, + "learning_rate": 0.0005568588713129409, + "loss": 4.344346523284912, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.027895759791135788, + "learning_rate": 0.0005567203972146906, + "loss": 4.365072727203369, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.026465769857168198, + "learning_rate": 0.000556581718698499, + "loss": 4.334959506988525, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.03191852569580078, + "learning_rate": 0.000556442835876097, + "loss": 4.361069679260254, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.03535444661974907, + "learning_rate": 0.0005563037488593799, + "loss": 4.349765777587891, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03451581299304962, + "learning_rate": 0.0005561644577604074, + "loss": 4.343142509460449, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029810845851898193, + "learning_rate": 0.000556024962691404, + "loss": 4.307485580444336, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.027523217722773552, + "learning_rate": 0.0005558852637647584, + "loss": 4.366406440734863, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03043208085000515, + "learning_rate": 0.0005557453610930235, + "loss": 4.356929779052734, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.03058922104537487, + "learning_rate": 0.0005556052547889161, + "loss": 4.348942279815674, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.032052166759967804, + "learning_rate": 0.0005554649449653177, + "loss": 4.3496809005737305, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.03170694038271904, + "learning_rate": 0.0005553244317352731, + "loss": 4.315959930419922, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 4.4757890701293945, + "eval_runtime": 57.7451, + "eval_samples_per_second": 42.289, + "eval_steps_per_second": 1.333, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.029025450348854065, + "learning_rate": 0.0005551837152119915, + "loss": 4.288647174835205, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.029403962194919586, + "learning_rate": 0.0005550427955088455, + "loss": 4.288555145263672, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.03265019878745079, + "learning_rate": 0.0005549016727393715, + "loss": 4.292625904083252, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.0387202650308609, + "learning_rate": 0.0005547603470172697, + "loss": 4.238926887512207, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.04554206505417824, + "learning_rate": 0.0005546188184564036, + "loss": 4.29966402053833, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.036231476813554764, + "learning_rate": 0.0005544770871708004, + "loss": 4.304398536682129, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.038703009486198425, + "learning_rate": 0.0005543351532746501, + "loss": 4.2945356369018555, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03224619850516319, + "learning_rate": 0.0005541930168823067, + "loss": 4.261351585388184, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.029722856357693672, + "learning_rate": 0.0005540506781082864, + "loss": 4.303143501281738, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.02764045260846615, + "learning_rate": 0.0005539081370672695, + "loss": 4.243679523468018, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.0312200877815485, + "learning_rate": 0.0005537653938740984, + "loss": 4.279904365539551, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.027349065989255905, + "learning_rate": 0.0005536224486437788, + "loss": 4.270228385925293, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.030385233461856842, + "learning_rate": 0.0005534793014914793, + "loss": 4.301590442657471, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.03292274847626686, + "learning_rate": 0.0005533359525325307, + "loss": 4.301828384399414, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.03265977278351784, + "learning_rate": 0.0005531924018824266, + "loss": 4.286200523376465, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.03177836537361145, + "learning_rate": 0.0005530486496568236, + "loss": 4.296319007873535, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.028879934921860695, + "learning_rate": 0.0005529046959715397, + "loss": 4.261706829071045, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.02855812944471836, + "learning_rate": 0.0005527605409425562, + "loss": 4.273903846740723, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03037901036441326, + "learning_rate": 0.0005526161846860161, + "loss": 4.248080253601074, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.029279105365276337, + "learning_rate": 0.0005524716273182245, + "loss": 4.281311988830566, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.02592580020427704, + "learning_rate": 0.0005523268689556489, + "loss": 4.204848289489746, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.021899603307247162, + "learning_rate": 0.0005521819097149183, + "loss": 4.232453346252441, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.019704442471265793, + "learning_rate": 0.0005520367497128238, + "loss": 4.253462314605713, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.019550519064068794, + "learning_rate": 0.0005518913890663182, + "loss": 4.307737350463867, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.020320113748311996, + "learning_rate": 0.0005517458278925161, + "loss": 4.257411956787109, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02118559181690216, + "learning_rate": 0.0005516000663086932, + "loss": 4.262089729309082, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.02050616219639778, + "learning_rate": 0.0005514541044322872, + "loss": 4.2490644454956055, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019793255254626274, + "learning_rate": 0.0005513079423808967, + "loss": 4.267400741577148, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.020760254934430122, + "learning_rate": 0.0005511615802722823, + "loss": 4.2532429695129395, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.019524799659848213, + "learning_rate": 0.0005510150182243647, + "loss": 4.267422199249268, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.020166637375950813, + "learning_rate": 0.0005508682563552269, + "loss": 4.214296817779541, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.01964394748210907, + "learning_rate": 0.0005507212947831119, + "loss": 4.252087116241455, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01690433733165264, + "learning_rate": 0.0005505741336264239, + "loss": 4.232336044311523, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.017682762816548347, + "learning_rate": 0.0005504267730037283, + "loss": 4.213962554931641, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01688133366405964, + "learning_rate": 0.0005502792130337508, + "loss": 4.199461460113525, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.019166061654686928, + "learning_rate": 0.0005501314538353776, + "loss": 4.228936672210693, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.024065518751740456, + "learning_rate": 0.000549983495527656, + "loss": 4.214227676391602, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02518288791179657, + "learning_rate": 0.000549835338229793, + "loss": 4.256412029266357, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.026093967258930206, + "learning_rate": 0.0005496869820611563, + "loss": 4.216564178466797, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.023775635287165642, + "learning_rate": 0.0005495384271412739, + "loss": 4.226430892944336, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.018810907378792763, + "learning_rate": 0.0005493896735898338, + "loss": 4.190013885498047, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.018974872305989265, + "learning_rate": 0.000549240721526684, + "loss": 4.246231555938721, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.02069833129644394, + "learning_rate": 0.0005490915710718325, + "loss": 4.198457717895508, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.02097306028008461, + "learning_rate": 0.0005489422223454471, + "loss": 4.212507247924805, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022165542468428612, + "learning_rate": 0.0005487926754678556, + "loss": 4.2043843269348145, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.019068919122219086, + "learning_rate": 0.0005486429305595448, + "loss": 4.197460651397705, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01701812446117401, + "learning_rate": 0.0005484929877411618, + "loss": 4.220056533813477, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01629726029932499, + "learning_rate": 0.0005483428471335127, + "loss": 4.214868545532227, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018661806359887123, + "learning_rate": 0.000548192508857563, + "loss": 4.227047920227051, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.02005891129374504, + "learning_rate": 0.0005480419730344377, + "loss": 4.197265625, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.02591128647327423, + "learning_rate": 0.0005478912397854208, + "loss": 4.211041450500488, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.032659754157066345, + "learning_rate": 0.0005477403092319552, + "loss": 4.230195045471191, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.04057808965444565, + "learning_rate": 0.0005475891814956429, + "loss": 4.223843097686768, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.04103454574942589, + "learning_rate": 0.0005474378566982449, + "loss": 4.197145462036133, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.03810729831457138, + "learning_rate": 0.0005472863349616809, + "loss": 4.237957000732422, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.028838301077485085, + "learning_rate": 0.0005471346164080292, + "loss": 4.206722259521484, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02436191774904728, + "learning_rate": 0.0005469827011595266, + "loss": 4.1983137130737305, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.022839387878775597, + "learning_rate": 0.0005468305893385686, + "loss": 4.20530366897583, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.02375260926783085, + "learning_rate": 0.0005466782810677088, + "loss": 4.213202476501465, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.02177741937339306, + "learning_rate": 0.0005465257764696593, + "loss": 4.200539588928223, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.02351468615233898, + "learning_rate": 0.0005463730756672903, + "loss": 4.198091506958008, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.023375794291496277, + "learning_rate": 0.0005462201787836301, + "loss": 4.193244934082031, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.02479805052280426, + "learning_rate": 0.0005460670859418651, + "loss": 4.2250871658325195, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.026894841343164444, + "learning_rate": 0.000545913797265339, + "loss": 4.241791248321533, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.030582403764128685, + "learning_rate": 0.0005457603128775541, + "loss": 4.203666687011719, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.032222259789705276, + "learning_rate": 0.0005456066329021699, + "loss": 4.209629535675049, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.03181060031056404, + "learning_rate": 0.0005454527574630037, + "loss": 4.200459957122803, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02687576413154602, + "learning_rate": 0.0005452986866840298, + "loss": 4.211819648742676, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.025988668203353882, + "learning_rate": 0.0005451444206893806, + "loss": 4.198040962219238, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.02937682718038559, + "learning_rate": 0.0005449899596033453, + "loss": 4.222935676574707, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.03043951839208603, + "learning_rate": 0.0005448353035503704, + "loss": 4.2031755447387695, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.030401621013879776, + "learning_rate": 0.0005446804526550596, + "loss": 4.198380470275879, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 4.368444919586182, + "eval_runtime": 58.3009, + "eval_samples_per_second": 41.886, + "eval_steps_per_second": 1.321, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.03461585193872452, + "learning_rate": 0.0005445254070421732, + "loss": 4.147369384765625, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.039084456861019135, + "learning_rate": 0.0005443701668366288, + "loss": 4.124129295349121, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.04021848738193512, + "learning_rate": 0.0005442147321635007, + "loss": 4.127362251281738, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.041000768542289734, + "learning_rate": 0.00054405910314802, + "loss": 4.1432061195373535, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.04008613899350166, + "learning_rate": 0.0005439032799155737, + "loss": 4.169931888580322, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.039394307881593704, + "learning_rate": 0.000543747262591706, + "loss": 4.155813217163086, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.04374868795275688, + "learning_rate": 0.0005435910513021172, + "loss": 4.159891128540039, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0443582646548748, + "learning_rate": 0.0005434346461726641, + "loss": 4.157362937927246, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.04601646214723587, + "learning_rate": 0.0005432780473293593, + "loss": 4.152097225189209, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.04327242821455002, + "learning_rate": 0.0005431212548983718, + "loss": 4.170074462890625, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.0413462370634079, + "learning_rate": 0.0005429642690060262, + "loss": 4.156710147857666, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.036855220794677734, + "learning_rate": 0.0005428070897788034, + "loss": 4.138053894042969, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.03781217709183693, + "learning_rate": 0.0005426497173433398, + "loss": 4.172306060791016, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.03886301815509796, + "learning_rate": 0.0005424921518264275, + "loss": 4.135544776916504, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.03730994835495949, + "learning_rate": 0.0005423343933550143, + "loss": 4.171654224395752, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.03271950036287308, + "learning_rate": 0.0005421764420562032, + "loss": 4.169906139373779, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.03292673081159592, + "learning_rate": 0.0005420182980572527, + "loss": 4.163660049438477, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.029366906732320786, + "learning_rate": 0.0005418599614855768, + "loss": 4.161055088043213, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.023805437609553337, + "learning_rate": 0.0005417014324687442, + "loss": 4.172390937805176, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.023178137838840485, + "learning_rate": 0.000541542711134479, + "loss": 4.146788597106934, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.02403450198471546, + "learning_rate": 0.0005413837976106599, + "loss": 4.124364852905273, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.025009717792272568, + "learning_rate": 0.0005412246920253208, + "loss": 4.12275505065918, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.023431289941072464, + "learning_rate": 0.0005410653945066503, + "loss": 4.134739398956299, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.023803815245628357, + "learning_rate": 0.0005409059051829914, + "loss": 4.130463123321533, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.021219994872808456, + "learning_rate": 0.0005407462241828417, + "loss": 4.124577522277832, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.01899666152894497, + "learning_rate": 0.0005405863516348535, + "loss": 4.167086124420166, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.01889493502676487, + "learning_rate": 0.000540426287667833, + "loss": 4.152744293212891, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.019546914845705032, + "learning_rate": 0.000540266032410741, + "loss": 4.122655868530273, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02085448056459427, + "learning_rate": 0.0005401055859926923, + "loss": 4.141949653625488, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.019959403201937675, + "learning_rate": 0.0005399449485429555, + "loss": 4.131907939910889, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.01885392889380455, + "learning_rate": 0.0005397841201909535, + "loss": 4.118953227996826, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.018216131255030632, + "learning_rate": 0.0005396231010662627, + "loss": 4.096459865570068, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.019385207444429398, + "learning_rate": 0.0005394618912986132, + "loss": 4.111050128936768, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.018676603212952614, + "learning_rate": 0.0005393004910178889, + "loss": 4.156215667724609, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.01876339502632618, + "learning_rate": 0.0005391389003541271, + "loss": 4.1107282638549805, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.01831623539328575, + "learning_rate": 0.0005389771194375185, + "loss": 4.154116630554199, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.018896661698818207, + "learning_rate": 0.0005388151483984071, + "loss": 4.093320846557617, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.019582601264119148, + "learning_rate": 0.0005386529873672898, + "loss": 4.106500625610352, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.020313601940870285, + "learning_rate": 0.0005384906364748172, + "loss": 4.071961879730225, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.020385954529047012, + "learning_rate": 0.000538328095851792, + "loss": 4.129006862640381, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02231123298406601, + "learning_rate": 0.0005381653656291707, + "loss": 4.067756652832031, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.022957179695367813, + "learning_rate": 0.0005380024459380619, + "loss": 4.088230133056641, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021438462659716606, + "learning_rate": 0.000537839336909727, + "loss": 4.128036022186279, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01893841102719307, + "learning_rate": 0.00053767603867558, + "loss": 4.155407905578613, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.018539676442742348, + "learning_rate": 0.0005375125513671874, + "loss": 4.101880073547363, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.01881185919046402, + "learning_rate": 0.0005373488751162681, + "loss": 4.118906021118164, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.021657396107912064, + "learning_rate": 0.0005371850100546929, + "loss": 4.091935157775879, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.023293187841773033, + "learning_rate": 0.000537020956314485, + "loss": 4.105803489685059, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.021306317299604416, + "learning_rate": 0.0005368567140278196, + "loss": 4.0610809326171875, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.022078553214669228, + "learning_rate": 0.0005366922833270235, + "loss": 4.09867525100708, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.02266664244234562, + "learning_rate": 0.0005365276643445757, + "loss": 4.1027326583862305, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.02091757394373417, + "learning_rate": 0.0005363628572131069, + "loss": 4.080266952514648, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.020374659448862076, + "learning_rate": 0.0005361978620653991, + "loss": 4.088479042053223, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.02294750325381756, + "learning_rate": 0.0005360326790343858, + "loss": 4.084364891052246, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.02337353676557541, + "learning_rate": 0.000535867308253152, + "loss": 4.125696182250977, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.022643696516752243, + "learning_rate": 0.0005357017498549341, + "loss": 4.097160339355469, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.021080605685710907, + "learning_rate": 0.0005355360039731196, + "loss": 4.102363109588623, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.02269427664577961, + "learning_rate": 0.0005353700707412466, + "loss": 4.089844703674316, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.02195413038134575, + "learning_rate": 0.0005352039502930048, + "loss": 4.132293701171875, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.01805616170167923, + "learning_rate": 0.0005350376427622343, + "loss": 4.0842814445495605, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.018048783764243126, + "learning_rate": 0.0005348711482829261, + "loss": 4.109591007232666, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.0199508648365736, + "learning_rate": 0.000534704466989222, + "loss": 4.0908074378967285, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.020186785608530045, + "learning_rate": 0.0005345375990154138, + "loss": 4.077959060668945, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.022526193410158157, + "learning_rate": 0.0005343705444959441, + "loss": 4.109979629516602, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.025224441662430763, + "learning_rate": 0.000534203303565406, + "loss": 4.047533988952637, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.027292724698781967, + "learning_rate": 0.000534035876358542, + "loss": 4.09566593170166, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.025669090449810028, + "learning_rate": 0.0005338682630102454, + "loss": 4.0606160163879395, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02791150100529194, + "learning_rate": 0.0005337004636555595, + "loss": 4.097081184387207, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.03294133022427559, + "learning_rate": 0.0005335324784296768, + "loss": 4.121442794799805, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.027322066947817802, + "learning_rate": 0.0005333643074679403, + "loss": 4.1188740730285645, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.024866372346878052, + "learning_rate": 0.000533195950905842, + "loss": 4.082042694091797, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.02901027724146843, + "learning_rate": 0.0005330274088790239, + "loss": 4.110199451446533, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 4.283970355987549, + "eval_runtime": 58.4003, + "eval_samples_per_second": 41.815, + "eval_steps_per_second": 1.318, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.028958842158317566, + "learning_rate": 0.0005328586815232773, + "loss": 4.040733337402344, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.03491180017590523, + "learning_rate": 0.0005326897689745428, + "loss": 4.0473432540893555, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.04025818780064583, + "learning_rate": 0.0005325206713689101, + "loss": 4.021976470947266, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03998899459838867, + "learning_rate": 0.0005323513888426181, + "loss": 4.04901123046875, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.042165111750364304, + "learning_rate": 0.0005321819215320546, + "loss": 4.050230979919434, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.03511597216129303, + "learning_rate": 0.0005320122695737565, + "loss": 4.0989990234375, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.04064979776740074, + "learning_rate": 0.0005318424331044094, + "loss": 4.051448822021484, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.047536592930555344, + "learning_rate": 0.0005316724122608473, + "loss": 4.064278602600098, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.05592669919133186, + "learning_rate": 0.0005315022071800527, + "loss": 4.045516014099121, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.04262633994221687, + "learning_rate": 0.0005313318179991574, + "loss": 4.060666084289551, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03997781500220299, + "learning_rate": 0.0005311612448554402, + "loss": 4.065315246582031, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.038152433931827545, + "learning_rate": 0.0005309904878863293, + "loss": 4.069721221923828, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.0367623008787632, + "learning_rate": 0.0005308195472294001, + "loss": 4.080986976623535, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.0358598455786705, + "learning_rate": 0.0005306484230223769, + "loss": 4.065290927886963, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.039188165217638016, + "learning_rate": 0.000530477115403131, + "loss": 4.024333477020264, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.036839909851551056, + "learning_rate": 0.0005303056245096817, + "loss": 4.067532539367676, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.03456985577940941, + "learning_rate": 0.0005301339504801967, + "loss": 4.048443794250488, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.039487943053245544, + "learning_rate": 0.0005299620934529902, + "loss": 4.0585737228393555, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.03415827825665474, + "learning_rate": 0.0005297900535665244, + "loss": 4.06104850769043, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.03209620714187622, + "learning_rate": 0.0005296178309594087, + "loss": 4.009165287017822, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.028947360813617706, + "learning_rate": 0.0005294454257704, + "loss": 4.067384243011475, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026836825534701347, + "learning_rate": 0.0005292728381384018, + "loss": 4.046419620513916, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024095216765999794, + "learning_rate": 0.0005291000682024647, + "loss": 4.019008636474609, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.020965851843357086, + "learning_rate": 0.0005289271161017867, + "loss": 4.020613670349121, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.02250855043530464, + "learning_rate": 0.0005287539819757119, + "loss": 4.031634330749512, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.021385207772254944, + "learning_rate": 0.0005285806659637315, + "loss": 4.086246013641357, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.021104589104652405, + "learning_rate": 0.000528407168205483, + "loss": 4.033970355987549, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.019584037363529205, + "learning_rate": 0.0005282334888407503, + "loss": 4.025118827819824, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.02107258513569832, + "learning_rate": 0.000528059628009464, + "loss": 4.020823955535889, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.020438389852643013, + "learning_rate": 0.0005278855858517003, + "loss": 4.016323566436768, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.01881488971412182, + "learning_rate": 0.0005277113625076822, + "loss": 3.9991841316223145, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.01763618178665638, + "learning_rate": 0.000527536958117778, + "loss": 4.0062150955200195, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.017308862879872322, + "learning_rate": 0.0005273623728225022, + "loss": 3.946723461151123, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01706845872104168, + "learning_rate": 0.0005271876067625153, + "loss": 4.056562900543213, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.018664294853806496, + "learning_rate": 0.0005270126600786228, + "loss": 4.024652481079102, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.018523404374718666, + "learning_rate": 0.0005268375329117762, + "loss": 3.962064743041992, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.017408935353159904, + "learning_rate": 0.0005266622254030724, + "loss": 4.008697509765625, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.01588570512831211, + "learning_rate": 0.0005264867376937534, + "loss": 4.011664390563965, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.015084310434758663, + "learning_rate": 0.0005263110699252065, + "loss": 3.996224880218506, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01637054979801178, + "learning_rate": 0.000526135222238964, + "loss": 4.015972137451172, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.01673802360892296, + "learning_rate": 0.0005259591947767032, + "loss": 3.975414276123047, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.016544019803404808, + "learning_rate": 0.0005257829876802464, + "loss": 4.012955188751221, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.017455460503697395, + "learning_rate": 0.0005256066010915603, + "loss": 4.010180473327637, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.017625460401177406, + "learning_rate": 0.0005254300351527566, + "loss": 4.001679420471191, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.01732795126736164, + "learning_rate": 0.000525253290006091, + "loss": 4.006107330322266, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.016654646024107933, + "learning_rate": 0.0005250763657939641, + "loss": 4.045180797576904, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.015085658989846706, + "learning_rate": 0.0005248992626589203, + "loss": 3.994246482849121, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.015112568624317646, + "learning_rate": 0.0005247219807436489, + "loss": 4.004963397979736, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.014882124960422516, + "learning_rate": 0.000524544520190982, + "loss": 4.029565811157227, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.016012471169233322, + "learning_rate": 0.000524366881143897, + "loss": 4.018208026885986, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.015761924907565117, + "learning_rate": 0.0005241890637455141, + "loss": 3.9911205768585205, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.017770355567336082, + "learning_rate": 0.0005240110681390978, + "loss": 4.028733253479004, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01959751546382904, + "learning_rate": 0.0005238328944680558, + "loss": 4.003322124481201, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.018713824450969696, + "learning_rate": 0.0005236545428759394, + "loss": 3.991718292236328, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.01992208883166313, + "learning_rate": 0.0005234760135064434, + "loss": 4.002507209777832, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.021671291440725327, + "learning_rate": 0.0005232973065034055, + "loss": 4.021344184875488, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024282798171043396, + "learning_rate": 0.0005231184220108069, + "loss": 3.9899158477783203, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.023294344544410706, + "learning_rate": 0.0005229393601727713, + "loss": 4.046627044677734, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.02029707282781601, + "learning_rate": 0.000522760121133566, + "loss": 4.022454261779785, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.02151239849627018, + "learning_rate": 0.0005225807050376004, + "loss": 4.005476951599121, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.02320881187915802, + "learning_rate": 0.0005224011120294269, + "loss": 4.012726783752441, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.026484191417694092, + "learning_rate": 0.0005222213422537401, + "loss": 3.9904391765594482, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.026279661804437637, + "learning_rate": 0.0005220413958553777, + "loss": 3.9738848209381104, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.024299729615449905, + "learning_rate": 0.0005218612729793188, + "loss": 4.003316402435303, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.02337595447897911, + "learning_rate": 0.0005216809737706856, + "loss": 3.954984188079834, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.02416428178548813, + "learning_rate": 0.0005215004983747415, + "loss": 4.0317063331604, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.022488107904791832, + "learning_rate": 0.0005213198469368925, + "loss": 3.9736647605895996, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.019700225442647934, + "learning_rate": 0.0005211390196026861, + "loss": 3.979867458343506, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.020382333546876907, + "learning_rate": 0.0005209580165178117, + "loss": 3.995251178741455, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.020149298012256622, + "learning_rate": 0.0005207768378281003, + "loss": 3.9987268447875977, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.019790509715676308, + "learning_rate": 0.0005205954836795241, + "loss": 3.9838991165161133, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.022096451371908188, + "learning_rate": 0.000520413954218197, + "loss": 3.995800256729126, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.214347839355469, + "eval_runtime": 58.8872, + "eval_samples_per_second": 41.469, + "eval_steps_per_second": 1.308, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02203402668237686, + "learning_rate": 0.0005202322495903739, + "loss": 3.915332078933716, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.02336675301194191, + "learning_rate": 0.0005200503699424509, + "loss": 3.914801597595215, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.027339637279510498, + "learning_rate": 0.0005198683154209654, + "loss": 3.9277780055999756, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.026872653514146805, + "learning_rate": 0.0005196860861725952, + "loss": 3.9160094261169434, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.026558255776762962, + "learning_rate": 0.0005195036823441592, + "loss": 3.9084300994873047, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.02644367143511772, + "learning_rate": 0.0005193211040826169, + "loss": 3.9328718185424805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.024677276611328125, + "learning_rate": 0.0005191383515350681, + "loss": 3.88993501663208, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.023194244131445885, + "learning_rate": 0.0005189554248487536, + "loss": 3.923567056655884, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.024653147906064987, + "learning_rate": 0.0005187723241710539, + "loss": 3.8740999698638916, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.026041202247142792, + "learning_rate": 0.0005185890496494899, + "loss": 3.9542148113250732, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.029586758464574814, + "learning_rate": 0.0005184056014317225, + "loss": 3.9241347312927246, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.0325787179172039, + "learning_rate": 0.0005182219796655528, + "loss": 3.9656593799591064, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.03183380514383316, + "learning_rate": 0.0005180381844989214, + "loss": 3.950648307800293, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.03434471786022186, + "learning_rate": 0.0005178542160799088, + "loss": 3.9590439796447754, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.030099432915449142, + "learning_rate": 0.0005176700745567353, + "loss": 3.9518113136291504, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.025950564071536064, + "learning_rate": 0.0005174857600777599, + "loss": 3.8908939361572266, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.029959935694932938, + "learning_rate": 0.0005173012727914816, + "loss": 3.945037364959717, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.029551763087511063, + "learning_rate": 0.0005171166128465389, + "loss": 3.923959732055664, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.03538523241877556, + "learning_rate": 0.0005169317803917083, + "loss": 3.9395527839660645, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.037924155592918396, + "learning_rate": 0.0005167467755759065, + "loss": 3.945115089416504, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.036773014813661575, + "learning_rate": 0.0005165615985481881, + "loss": 3.9484646320343018, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.04125838354229927, + "learning_rate": 0.0005163762494577474, + "loss": 4.000699043273926, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.044125135987997055, + "learning_rate": 0.0005161907284539165, + "loss": 3.9828028678894043, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.040853384882211685, + "learning_rate": 0.0005160050356861663, + "loss": 3.957505702972412, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.03501008450984955, + "learning_rate": 0.0005158191713041063, + "loss": 3.9360849857330322, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.04082556813955307, + "learning_rate": 0.0005156331354574838, + "loss": 3.95941424369812, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.045432962477207184, + "learning_rate": 0.0005154469282961847, + "loss": 3.9666922092437744, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.04048244655132294, + "learning_rate": 0.0005152605499702328, + "loss": 3.984555244445801, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.031016314402222633, + "learning_rate": 0.0005150740006297896, + "loss": 3.945204257965088, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.03210771828889847, + "learning_rate": 0.0005148872804251548, + "loss": 3.976186752319336, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.03130023553967476, + "learning_rate": 0.000514700389506765, + "loss": 3.915727138519287, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.031193193048238754, + "learning_rate": 0.0005145133280251954, + "loss": 3.961866855621338, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.027518954128026962, + "learning_rate": 0.0005143260961311578, + "loss": 3.9742956161499023, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.02751155197620392, + "learning_rate": 0.0005141386939755014, + "loss": 3.949662208557129, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.026033353060483932, + "learning_rate": 0.0005139511217092129, + "loss": 3.9327926635742188, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.02579621411859989, + "learning_rate": 0.000513763379483416, + "loss": 3.9334418773651123, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.023009559139609337, + "learning_rate": 0.0005135754674493709, + "loss": 3.969707489013672, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.02196447364985943, + "learning_rate": 0.0005133873857584753, + "loss": 3.9084434509277344, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.02206484042108059, + "learning_rate": 0.0005131991345622629, + "loss": 3.922117233276367, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.023360205814242363, + "learning_rate": 0.0005130107140124046, + "loss": 3.965158224105835, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02162669226527214, + "learning_rate": 0.0005128221242607072, + "loss": 3.9519729614257812, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.019719846546649933, + "learning_rate": 0.0005126333654591141, + "loss": 3.9195961952209473, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.018988260999321938, + "learning_rate": 0.0005124444377597049, + "loss": 3.937926769256592, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.019278401508927345, + "learning_rate": 0.0005122553413146951, + "loss": 3.9126808643341064, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019101005047559738, + "learning_rate": 0.0005120660762764366, + "loss": 3.8993749618530273, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.019156700000166893, + "learning_rate": 0.0005118766427974166, + "loss": 3.9531068801879883, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019944407045841217, + "learning_rate": 0.0005116870410302582, + "loss": 3.9645180702209473, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.01826174184679985, + "learning_rate": 0.0005114972711277202, + "loss": 3.948477268218994, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.0191368218511343, + "learning_rate": 0.0005113073332426968, + "loss": 3.944648027420044, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.017768343910574913, + "learning_rate": 0.0005111172275282174, + "loss": 3.9584474563598633, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.01955590397119522, + "learning_rate": 0.0005109269541374469, + "loss": 3.8936800956726074, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.02100536972284317, + "learning_rate": 0.000510736513223685, + "loss": 3.93919038772583, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.020873935893177986, + "learning_rate": 0.0005105459049403665, + "loss": 3.929244041442871, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.021876046434044838, + "learning_rate": 0.0005103551294410614, + "loss": 3.9264731407165527, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.021726684644818306, + "learning_rate": 0.0005101641868794736, + "loss": 3.923135280609131, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.019308188930153847, + "learning_rate": 0.0005099730774094423, + "loss": 3.9062135219573975, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01917017623782158, + "learning_rate": 0.0005097818011849408, + "loss": 3.898810386657715, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017933931201696396, + "learning_rate": 0.0005095903583600772, + "loss": 3.8899734020233154, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.015131795778870583, + "learning_rate": 0.0005093987490890933, + "loss": 3.9452900886535645, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.017190825194120407, + "learning_rate": 0.0005092069735263651, + "loss": 3.9124746322631836, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.01618487387895584, + "learning_rate": 0.000509015031826403, + "loss": 3.91654109954834, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.014920198358595371, + "learning_rate": 0.0005088229241438509, + "loss": 3.9144535064697266, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01615205965936184, + "learning_rate": 0.0005086306506334864, + "loss": 3.917003631591797, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.015370192006230354, + "learning_rate": 0.0005084382114502208, + "loss": 3.8910603523254395, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.015290996059775352, + "learning_rate": 0.0005082456067490988, + "loss": 3.9122695922851562, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.015913106501102448, + "learning_rate": 0.0005080528366852987, + "loss": 3.9304556846618652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.01639600843191147, + "learning_rate": 0.0005078599014141318, + "loss": 3.8974502086639404, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.017010601237416267, + "learning_rate": 0.0005076668010910425, + "loss": 3.9723148345947266, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.017778996378183365, + "learning_rate": 0.0005074735358716083, + "loss": 3.8905391693115234, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.020029611885547638, + "learning_rate": 0.0005072801059115394, + "loss": 3.934976100921631, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.01931857503950596, + "learning_rate": 0.000507086511366679, + "loss": 3.9152774810791016, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019750485196709633, + "learning_rate": 0.0005068927523930027, + "loss": 3.877650499343872, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.161914825439453, + "eval_runtime": 58.5217, + "eval_samples_per_second": 41.728, + "eval_steps_per_second": 1.316, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02073049545288086, + "learning_rate": 0.0005066988291466184, + "loss": 3.8331246376037598, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02272389829158783, + "learning_rate": 0.0005065047417837669, + "loss": 3.8847155570983887, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02139078453183174, + "learning_rate": 0.0005063104904608206, + "loss": 3.8276169300079346, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02043229155242443, + "learning_rate": 0.0005061160753342843, + "loss": 3.8804001808166504, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.02345564216375351, + "learning_rate": 0.0005059214965607948, + "loss": 3.851609468460083, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.023448185995221138, + "learning_rate": 0.0005057267542971207, + "loss": 3.8446507453918457, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.024826066568493843, + "learning_rate": 0.0005055318487001624, + "loss": 3.86712384223938, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.02491980977356434, + "learning_rate": 0.0005053367799269518, + "loss": 3.8374528884887695, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.023002834990620613, + "learning_rate": 0.0005051415481346522, + "loss": 3.783214569091797, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.02123952843248844, + "learning_rate": 0.0005049461534805585, + "loss": 3.8463428020477295, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.01981225423514843, + "learning_rate": 0.0005047505961220964, + "loss": 3.8307082653045654, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01915367692708969, + "learning_rate": 0.0005045548762168231, + "loss": 3.8190226554870605, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.01890859752893448, + "learning_rate": 0.0005043589939224266, + "loss": 3.8428220748901367, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.019150376319885254, + "learning_rate": 0.0005041629493967257, + "loss": 3.830247402191162, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.020099453628063202, + "learning_rate": 0.0005039667427976697, + "loss": 3.838806629180908, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.021136781200766563, + "learning_rate": 0.0005037703742833391, + "loss": 3.8344273567199707, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.022857150062918663, + "learning_rate": 0.0005035738440119442, + "loss": 3.8317880630493164, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.024891739711165428, + "learning_rate": 0.0005033771521418261, + "loss": 3.837855577468872, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.02595464698970318, + "learning_rate": 0.0005031802988314557, + "loss": 3.846050262451172, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.02722409926354885, + "learning_rate": 0.0005029832842394342, + "loss": 3.8164305686950684, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.024486860260367393, + "learning_rate": 0.0005027861085244927, + "loss": 3.8099875450134277, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.021551169455051422, + "learning_rate": 0.0005025887718454922, + "loss": 3.825925827026367, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.02121506817638874, + "learning_rate": 0.0005023912743614232, + "loss": 3.870727777481079, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.01948230341076851, + "learning_rate": 0.0005021936162314058, + "loss": 3.8035314083099365, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.01896044984459877, + "learning_rate": 0.0005019957976146898, + "loss": 3.84665584564209, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.020620400086045265, + "learning_rate": 0.0005017978186706538, + "loss": 3.8338310718536377, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.020855262875556946, + "learning_rate": 0.0005015996795588062, + "loss": 3.8781003952026367, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.02153870463371277, + "learning_rate": 0.0005014013804387838, + "loss": 3.8783490657806396, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.023751024156808853, + "learning_rate": 0.0005012029214703529, + "loss": 3.846198558807373, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.023987066000699997, + "learning_rate": 0.000501004302813408, + "loss": 3.834296226501465, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.023626182228326797, + "learning_rate": 0.0005008055246279727, + "loss": 3.8591227531433105, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.025707192718982697, + "learning_rate": 0.000500606587074199, + "loss": 3.8766489028930664, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.027723029255867004, + "learning_rate": 0.0005004074903123674, + "loss": 3.88264799118042, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.0278895553201437, + "learning_rate": 0.0005002082345028864, + "loss": 3.8932995796203613, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.025622662156820297, + "learning_rate": 0.0005000088198062929, + "loss": 3.8734612464904785, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.02639615908265114, + "learning_rate": 0.0004998092463832516, + "loss": 3.786813735961914, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.026765599846839905, + "learning_rate": 0.0004996095143945554, + "loss": 3.853966474533081, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02823319099843502, + "learning_rate": 0.0004994096240011246, + "loss": 3.8321340084075928, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.023394955322146416, + "learning_rate": 0.0004992095753640072, + "loss": 3.8527331352233887, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.023874662816524506, + "learning_rate": 0.0004990093686443791, + "loss": 3.8822879791259766, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.02301311306655407, + "learning_rate": 0.000498809004003543, + "loss": 3.846186876296997, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.022630592808127403, + "learning_rate": 0.0004986084816029291, + "loss": 3.867992639541626, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.02424517087638378, + "learning_rate": 0.0004984078016040946, + "loss": 3.867293357849121, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.024116624146699905, + "learning_rate": 0.000498206964168724, + "loss": 3.846282958984375, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.02187768742442131, + "learning_rate": 0.000498005969458628, + "loss": 3.8429484367370605, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.01879809983074665, + "learning_rate": 0.0004978048176357447, + "loss": 3.863215446472168, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.020426234230399132, + "learning_rate": 0.0004976035088621385, + "loss": 3.825230121612549, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01992812193930149, + "learning_rate": 0.0004974020433000003, + "loss": 3.851665496826172, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.020693328231573105, + "learning_rate": 0.000497200421111647, + "loss": 3.906999111175537, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.021137919276952744, + "learning_rate": 0.0004969986424595221, + "loss": 3.8369035720825195, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.023285407572984695, + "learning_rate": 0.0004967967075061949, + "loss": 3.8658390045166016, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.0219135619699955, + "learning_rate": 0.0004965946164143609, + "loss": 3.873072862625122, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.020101983100175858, + "learning_rate": 0.0004963923693468413, + "loss": 3.847805976867676, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.01875176839530468, + "learning_rate": 0.0004961899664665827, + "loss": 3.7976622581481934, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01913360133767128, + "learning_rate": 0.0004959874079366576, + "loss": 3.853851318359375, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.016863619908690453, + "learning_rate": 0.0004957846939202637, + "loss": 3.877378225326538, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.017802366986870766, + "learning_rate": 0.000495581824580724, + "loss": 3.861356019973755, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.01981814205646515, + "learning_rate": 0.0004953788000814866, + "loss": 3.88613224029541, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.017508897930383682, + "learning_rate": 0.000495175620586125, + "loss": 3.8202030658721924, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.019240548834204674, + "learning_rate": 0.0004949722862583368, + "loss": 3.853827714920044, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.022305598482489586, + "learning_rate": 0.000494768797261945, + "loss": 3.885148763656616, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.02531025931239128, + "learning_rate": 0.000494565153760897, + "loss": 3.8412041664123535, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.024474849924445152, + "learning_rate": 0.0004943613559192648, + "loss": 3.865644931793213, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.020430924370884895, + "learning_rate": 0.0004941574039012444, + "loss": 3.8649442195892334, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.01906431093811989, + "learning_rate": 0.0004939532978711566, + "loss": 3.8462204933166504, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.021118057891726494, + "learning_rate": 0.0004937490379934456, + "loss": 3.8652522563934326, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.020337693393230438, + "learning_rate": 0.00049354462443268, + "loss": 3.8506221771240234, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.019090797752141953, + "learning_rate": 0.0004933400573535521, + "loss": 3.881237506866455, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.020973797887563705, + "learning_rate": 0.000493135336920878, + "loss": 3.821274757385254, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.021984362974762917, + "learning_rate": 0.0004929304632995973, + "loss": 3.8793835639953613, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.0215684212744236, + "learning_rate": 0.0004927254366547727, + "loss": 3.8259663581848145, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.021673865616321564, + "learning_rate": 0.0004925202571515908, + "loss": 3.8505711555480957, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.123871326446533, + "eval_runtime": 58.8357, + "eval_samples_per_second": 41.505, + "eval_steps_per_second": 1.309, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.0203359667211771, + "learning_rate": 0.0004923149249553608, + "loss": 3.785353660583496, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.023197868838906288, + "learning_rate": 0.0004921094402315152, + "loss": 3.7479515075683594, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.02314298041164875, + "learning_rate": 0.0004919038031456093, + "loss": 3.7697315216064453, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.0256908368319273, + "learning_rate": 0.0004916980138633213, + "loss": 3.79056978225708, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.026193121448159218, + "learning_rate": 0.0004914920725504518, + "loss": 3.7722532749176025, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.027120303362607956, + "learning_rate": 0.000491285979372924, + "loss": 3.7991981506347656, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.0266975536942482, + "learning_rate": 0.0004910797344967834, + "loss": 3.778301477432251, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.02806396596133709, + "learning_rate": 0.000490873338088198, + "loss": 3.7578704357147217, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.02893839031457901, + "learning_rate": 0.0004906667903134573, + "loss": 3.7401275634765625, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.031170370057225227, + "learning_rate": 0.0004904600913389735, + "loss": 3.780686616897583, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.03364266827702522, + "learning_rate": 0.0004902532413312799, + "loss": 3.794442653656006, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.030121006071567535, + "learning_rate": 0.0004900462404570321, + "loss": 3.779569149017334, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.03296204283833504, + "learning_rate": 0.0004898390888830067, + "loss": 3.7960336208343506, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.03381749242544174, + "learning_rate": 0.0004896317867761022, + "loss": 3.7959280014038086, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.035383883863687515, + "learning_rate": 0.000489424334303338, + "loss": 3.8008625507354736, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.03759761154651642, + "learning_rate": 0.0004892167316318548, + "loss": 3.7997028827667236, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.030651796609163284, + "learning_rate": 0.0004890089789289147, + "loss": 3.81191349029541, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.0262477844953537, + "learning_rate": 0.0004888010763618997, + "loss": 3.8046622276306152, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.030398324131965637, + "learning_rate": 0.0004885930240983136, + "loss": 3.7903637886047363, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.030515259131789207, + "learning_rate": 0.0004883848223057801, + "loss": 3.8111231327056885, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.031686048954725266, + "learning_rate": 0.00048817647115204386, + "loss": 3.781445026397705, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.025968654081225395, + "learning_rate": 0.00048796797080496936, + "loss": 3.788196563720703, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.025340190157294273, + "learning_rate": 0.00048775932143254176, + "loss": 3.763331413269043, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.024738596752285957, + "learning_rate": 0.0004875505232028661, + "loss": 3.8005969524383545, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.021990075707435608, + "learning_rate": 0.0004873415762841672, + "loss": 3.795844078063965, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.02191784791648388, + "learning_rate": 0.00048713248084478986, + "loss": 3.8062384128570557, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.023678259924054146, + "learning_rate": 0.0004869232370531986, + "loss": 3.7760679721832275, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.024137988686561584, + "learning_rate": 0.0004867138450779773, + "loss": 3.7917492389678955, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.02673223614692688, + "learning_rate": 0.0004865043050878293, + "loss": 3.8006649017333984, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.02395169623196125, + "learning_rate": 0.0004862946172515771, + "loss": 3.7902212142944336, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.02410844899713993, + "learning_rate": 0.0004860847817381625, + "loss": 3.8073291778564453, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.028109831735491753, + "learning_rate": 0.0004858747987166463, + "loss": 3.80116605758667, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.025538964197039604, + "learning_rate": 0.00048566466835620774, + "loss": 3.791151523590088, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.02393815852701664, + "learning_rate": 0.0004854543908261453, + "loss": 3.7766106128692627, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.024659741669893265, + "learning_rate": 0.0004852439662958756, + "loss": 3.789135217666626, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.02668776921927929, + "learning_rate": 0.000485033394934934, + "loss": 3.797905445098877, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.0258641317486763, + "learning_rate": 0.000484822676912974, + "loss": 3.783323287963867, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.024846022948622704, + "learning_rate": 0.0004846118123997674, + "loss": 3.7976717948913574, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.024981874972581863, + "learning_rate": 0.00048440080156520387, + "loss": 3.818732738494873, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.02556881681084633, + "learning_rate": 0.0004841896445792909, + "loss": 3.7821896076202393, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.02558571845293045, + "learning_rate": 0.000483978341612154, + "loss": 3.8143680095672607, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.02361566200852394, + "learning_rate": 0.0004837668928340362, + "loss": 3.7710189819335938, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022570716217160225, + "learning_rate": 0.0004835552984152978, + "loss": 3.775564670562744, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.022478004917502403, + "learning_rate": 0.00048334355852641664, + "loss": 3.7920689582824707, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021241728216409683, + "learning_rate": 0.0004831316733379879, + "loss": 3.7857837677001953, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.018921511247754097, + "learning_rate": 0.0004829196430207235, + "loss": 3.7659268379211426, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01926037296652794, + "learning_rate": 0.00048270746774545237, + "loss": 3.7527129650115967, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.018743667751550674, + "learning_rate": 0.0004824951476831204, + "loss": 3.767991065979004, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.017273949459195137, + "learning_rate": 0.00048228268300479015, + "loss": 3.7834174633026123, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.017054244875907898, + "learning_rate": 0.0004820700738816403, + "loss": 3.7775752544403076, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017190931364893913, + "learning_rate": 0.0004818573204849664, + "loss": 3.7709648609161377, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01665034517645836, + "learning_rate": 0.00048164442298617987, + "loss": 3.7575321197509766, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.016932843253016472, + "learning_rate": 0.00048143138155680844, + "loss": 3.7922799587249756, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01702750287950039, + "learning_rate": 0.0004812181963684958, + "loss": 3.7729780673980713, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01766599342226982, + "learning_rate": 0.00048100486759300134, + "loss": 3.7876572608947754, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01785697042942047, + "learning_rate": 0.00048079139540220024, + "loss": 3.8131699562072754, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01830688677728176, + "learning_rate": 0.0004805777799680831, + "loss": 3.790778636932373, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017348438501358032, + "learning_rate": 0.0004803640214627561, + "loss": 3.7984330654144287, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017999377101659775, + "learning_rate": 0.0004801501200584406, + "loss": 3.7756147384643555, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.01801210641860962, + "learning_rate": 0.000479936075927473, + "loss": 3.789759874343872, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.018445398658514023, + "learning_rate": 0.000479721889242305, + "loss": 3.7683000564575195, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.017544370144605637, + "learning_rate": 0.0004795075601755027, + "loss": 3.7776038646698, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.019834233447909355, + "learning_rate": 0.00047929308889974737, + "loss": 3.7559571266174316, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01791626401245594, + "learning_rate": 0.0004790784755878346, + "loss": 3.7746729850769043, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.017638562247157097, + "learning_rate": 0.00047886372041267434, + "loss": 3.7933082580566406, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.018974050879478455, + "learning_rate": 0.0004786488235472909, + "loss": 3.795713424682617, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.018694303929805756, + "learning_rate": 0.0004784337851648231, + "loss": 3.7675955295562744, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.01788092777132988, + "learning_rate": 0.0004782186054385233, + "loss": 3.786083698272705, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.01902497000992298, + "learning_rate": 0.0004780032845417579, + "loss": 3.7976233959198, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.020558880642056465, + "learning_rate": 0.000477787822648007, + "loss": 3.7500152587890625, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.020093899220228195, + "learning_rate": 0.00047757221993086443, + "loss": 3.703448534011841, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.019833799451589584, + "learning_rate": 0.00047735647656403754, + "loss": 3.756059408187866, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.093987941741943, + "eval_runtime": 57.6894, + "eval_samples_per_second": 42.33, + "eval_steps_per_second": 1.335, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.022167526185512543, + "learning_rate": 0.0004771405927213467, + "loss": 3.696444511413574, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.027874166145920753, + "learning_rate": 0.0004769245685767255, + "loss": 3.713630199432373, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.030872784554958344, + "learning_rate": 0.00047670840430422113, + "loss": 3.678295612335205, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.03246859833598137, + "learning_rate": 0.0004764921000779928, + "loss": 3.7186388969421387, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.03667069226503372, + "learning_rate": 0.0004762756560723131, + "loss": 3.7457995414733887, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.036286015063524246, + "learning_rate": 0.0004760590724615671, + "loss": 3.6881775856018066, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.03413614630699158, + "learning_rate": 0.0004758423494202522, + "loss": 3.6999263763427734, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.03384552150964737, + "learning_rate": 0.0004756254871229782, + "loss": 3.6959803104400635, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.03227181360125542, + "learning_rate": 0.00047540848574446724, + "loss": 3.7182722091674805, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.036366235464811325, + "learning_rate": 0.0004751913454595535, + "loss": 3.7116851806640625, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.03573326766490936, + "learning_rate": 0.0004749740664431828, + "loss": 3.741865873336792, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.03482096269726753, + "learning_rate": 0.000474756648870413, + "loss": 3.7195520401000977, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.03232625126838684, + "learning_rate": 0.0004745390929164134, + "loss": 3.7443251609802246, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.03421664610505104, + "learning_rate": 0.0004743213987564651, + "loss": 3.6867361068725586, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.03585013002157211, + "learning_rate": 0.0004741035665659603, + "loss": 3.7492403984069824, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.029985329136252403, + "learning_rate": 0.00047388559652040236, + "loss": 3.705209493637085, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.02729589305818081, + "learning_rate": 0.0004736674887954059, + "loss": 3.711106538772583, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.029642216861248016, + "learning_rate": 0.0004734492435666963, + "loss": 3.7274727821350098, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.02816903032362461, + "learning_rate": 0.0004732308610101099, + "loss": 3.7181742191314697, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.027143625542521477, + "learning_rate": 0.00047301234130159346, + "loss": 3.719235897064209, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.025802744552493095, + "learning_rate": 0.00047279368461720445, + "loss": 3.771820545196533, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.026424532756209373, + "learning_rate": 0.0004725748911331106, + "loss": 3.7516651153564453, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.029677648097276688, + "learning_rate": 0.0004723559610255898, + "loss": 3.7058472633361816, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.031548649072647095, + "learning_rate": 0.0004721368944710302, + "loss": 3.756992816925049, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.027667395770549774, + "learning_rate": 0.00047191769164592974, + "loss": 3.742711067199707, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.027639545500278473, + "learning_rate": 0.000471698352726896, + "loss": 3.728883743286133, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.02950332500040531, + "learning_rate": 0.0004714788778906467, + "loss": 3.7160212993621826, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.026302607730031013, + "learning_rate": 0.0004712592673140084, + "loss": 3.735713481903076, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.022483112290501595, + "learning_rate": 0.00047103952117391764, + "loss": 3.702686309814453, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.021729422733187675, + "learning_rate": 0.00047081963964741986, + "loss": 3.7195816040039062, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.021159254014492035, + "learning_rate": 0.0004705996229116696, + "loss": 3.684321880340576, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.02233424223959446, + "learning_rate": 0.0004703794711439304, + "loss": 3.7233057022094727, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.02142048068344593, + "learning_rate": 0.0004701591845215744, + "loss": 3.7083680629730225, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.022331461310386658, + "learning_rate": 0.00046993876322208274, + "loss": 3.7408461570739746, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.020629866048693657, + "learning_rate": 0.0004697182074230448, + "loss": 3.759129047393799, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.02174685336649418, + "learning_rate": 0.0004694975173021586, + "loss": 3.72351336479187, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.021538110449910164, + "learning_rate": 0.0004692766930372299, + "loss": 3.716416597366333, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.020800625905394554, + "learning_rate": 0.00046905573480617285, + "loss": 3.726954936981201, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.02039588987827301, + "learning_rate": 0.00046883464278700966, + "loss": 3.7336390018463135, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.02107219770550728, + "learning_rate": 0.0004686134171578702, + "loss": 3.722311496734619, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.02068956382572651, + "learning_rate": 0.00046839205809699183, + "loss": 3.767139434814453, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.020190054550766945, + "learning_rate": 0.0004681705657827195, + "loss": 3.6960511207580566, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.018814411014318466, + "learning_rate": 0.0004679489403935058, + "loss": 3.70375919342041, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.017568519338965416, + "learning_rate": 0.00046772718210791005, + "loss": 3.756627321243286, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.018188195303082466, + "learning_rate": 0.00046750529110459903, + "loss": 3.6964683532714844, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.018925746902823448, + "learning_rate": 0.0004672832675623463, + "loss": 3.731172561645508, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.01690756157040596, + "learning_rate": 0.0004670611116600322, + "loss": 3.6955318450927734, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016300756484270096, + "learning_rate": 0.00046683882357664375, + "loss": 3.762632369995117, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.016405068337917328, + "learning_rate": 0.0004666164034912744, + "loss": 3.75130558013916, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.017687877640128136, + "learning_rate": 0.00046639385158312406, + "loss": 3.687530755996704, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.017524344846606255, + "learning_rate": 0.00046617116803149874, + "loss": 3.738790273666382, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.01714804768562317, + "learning_rate": 0.00046594835301581053, + "loss": 3.69827938079834, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018042324110865593, + "learning_rate": 0.00046572540671557745, + "loss": 3.71600341796875, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018648479133844376, + "learning_rate": 0.00046550232931042346, + "loss": 3.742591619491577, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.019198937341570854, + "learning_rate": 0.00046527912098007787, + "loss": 3.739820957183838, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.01796373911201954, + "learning_rate": 0.00046505578190437564, + "loss": 3.719025135040283, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01836620457470417, + "learning_rate": 0.0004648323122632569, + "loss": 3.741434097290039, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.02192627638578415, + "learning_rate": 0.0004646087122367673, + "loss": 3.7553882598876953, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.02147185243666172, + "learning_rate": 0.00046438498200505727, + "loss": 3.7705790996551514, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01855960488319397, + "learning_rate": 0.0004641611217483821, + "loss": 3.704925537109375, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.01925475336611271, + "learning_rate": 0.00046393713164710217, + "loss": 3.711491346359253, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019796503707766533, + "learning_rate": 0.00046371301188168204, + "loss": 3.7241992950439453, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.018511971458792686, + "learning_rate": 0.0004634887626326911, + "loss": 3.6785435676574707, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.018214019015431404, + "learning_rate": 0.00046326438408080293, + "loss": 3.7299962043762207, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.018912190571427345, + "learning_rate": 0.00046303987640679517, + "loss": 3.773698329925537, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.01939854398369789, + "learning_rate": 0.0004628152397915498, + "loss": 3.682321071624756, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.017859861254692078, + "learning_rate": 0.00046259047441605215, + "loss": 3.7254951000213623, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.019249707460403442, + "learning_rate": 0.0004623655804613919, + "loss": 3.7263433933258057, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.019221967086195946, + "learning_rate": 0.00046214055810876194, + "loss": 3.7229225635528564, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01871953159570694, + "learning_rate": 0.00046191540753945886, + "loss": 3.7248311042785645, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01747182011604309, + "learning_rate": 0.00046169012893488204, + "loss": 3.699913740158081, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.019630730152130127, + "learning_rate": 0.00046146472247653484, + "loss": 3.787783145904541, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.072091102600098, + "eval_runtime": 63.1589, + "eval_samples_per_second": 38.664, + "eval_steps_per_second": 1.219, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.022918934002518654, + "learning_rate": 0.0004612391883460228, + "loss": 3.666675567626953, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.0264838095754385, + "learning_rate": 0.00046101352672505493, + "loss": 3.6371259689331055, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.028463594615459442, + "learning_rate": 0.0004607877377954426, + "loss": 3.657130241394043, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.02900792844593525, + "learning_rate": 0.00046056182173909995, + "loss": 3.6365628242492676, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.028593607246875763, + "learning_rate": 0.0004603357787380435, + "loss": 3.647258996963501, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.02878790907561779, + "learning_rate": 0.0004601096089743919, + "loss": 3.682035446166992, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.02980183996260166, + "learning_rate": 0.0004598833126303661, + "loss": 3.623297691345215, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.03375725448131561, + "learning_rate": 0.00045965688988828884, + "loss": 3.6240830421447754, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.035045325756073, + "learning_rate": 0.00045943034093058507, + "loss": 3.6372249126434326, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.03519923985004425, + "learning_rate": 0.000459203665939781, + "loss": 3.6157641410827637, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.03468473628163338, + "learning_rate": 0.00045897686509850487, + "loss": 3.6553335189819336, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.034817393869161606, + "learning_rate": 0.0004587499385894856, + "loss": 3.653806686401367, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.03549059107899666, + "learning_rate": 0.000458522886595554, + "loss": 3.6730189323425293, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.028001410886645317, + "learning_rate": 0.0004582957092996418, + "loss": 3.662490129470825, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.02532520331442356, + "learning_rate": 0.0004580684068847817, + "loss": 3.629936456680298, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.028215918689966202, + "learning_rate": 0.0004578409795341069, + "loss": 3.6507444381713867, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.026823483407497406, + "learning_rate": 0.0004576134274308517, + "loss": 3.661811113357544, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.0281058382242918, + "learning_rate": 0.0004573857507583506, + "loss": 3.6805872917175293, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.02548503689467907, + "learning_rate": 0.00045715794970003874, + "loss": 3.6586413383483887, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.024795135483145714, + "learning_rate": 0.0004569300244394513, + "loss": 3.654989004135132, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.025501688942313194, + "learning_rate": 0.0004567019751602234, + "loss": 3.6915903091430664, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.02377370186150074, + "learning_rate": 0.00045647380204609034, + "loss": 3.660588264465332, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.022150425240397453, + "learning_rate": 0.00045624550528088704, + "loss": 3.64870023727417, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.02415974996984005, + "learning_rate": 0.00045601708504854803, + "loss": 3.6498191356658936, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.02140078693628311, + "learning_rate": 0.0004557885415331074, + "loss": 3.703489065170288, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.022121304646134377, + "learning_rate": 0.00045555987491869865, + "loss": 3.6606979370117188, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.02168990671634674, + "learning_rate": 0.00045533108538955416, + "loss": 3.6801178455352783, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.020228713750839233, + "learning_rate": 0.0004551021731300055, + "loss": 3.6707205772399902, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.018500352278351784, + "learning_rate": 0.0004548731383244834, + "loss": 3.647545576095581, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.020223816856741905, + "learning_rate": 0.0004546439811575169, + "loss": 3.644932985305786, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.019502121955156326, + "learning_rate": 0.00045441470181373375, + "loss": 3.62888240814209, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.01926770992577076, + "learning_rate": 0.0004541853004778603, + "loss": 3.675457715988159, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01888471283018589, + "learning_rate": 0.00045395577733472087, + "loss": 3.636434316635132, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.0201494712382555, + "learning_rate": 0.0004537261325692383, + "loss": 3.679147720336914, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.02178201824426651, + "learning_rate": 0.0004534963663664332, + "loss": 3.671905040740967, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.024073559790849686, + "learning_rate": 0.00045326647891142394, + "loss": 3.6701040267944336, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.02135559171438217, + "learning_rate": 0.0004530364703894268, + "loss": 3.648519992828369, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.019264448434114456, + "learning_rate": 0.0004528063409857554, + "loss": 3.6985023021698, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.01922295056283474, + "learning_rate": 0.0004525760908858209, + "loss": 3.651756763458252, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.01978331431746483, + "learning_rate": 0.00045234572027513174, + "loss": 3.679884910583496, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.01968599483370781, + "learning_rate": 0.00045211522933929317, + "loss": 3.6571812629699707, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.020741639658808708, + "learning_rate": 0.0004518846182640077, + "loss": 3.6505298614501953, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.01842017099261284, + "learning_rate": 0.0004516538872350744, + "loss": 3.6904940605163574, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.018004486337304115, + "learning_rate": 0.0004514230364383893, + "loss": 3.7282047271728516, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.018821008503437042, + "learning_rate": 0.0004511920660599444, + "loss": 3.668164014816284, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.018929854035377502, + "learning_rate": 0.0004509609762858287, + "loss": 3.6672980785369873, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.01979168877005577, + "learning_rate": 0.0004507297673022269, + "loss": 3.6684999465942383, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.01774384267628193, + "learning_rate": 0.0004504984392954199, + "loss": 3.6577749252319336, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01876000687479973, + "learning_rate": 0.00045026699245178444, + "loss": 3.6629252433776855, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.018197249621152878, + "learning_rate": 0.0004500354269577932, + "loss": 3.688983678817749, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.019496265798807144, + "learning_rate": 0.00044980374300001414, + "loss": 3.6602022647857666, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01825535111129284, + "learning_rate": 0.00044957194076511095, + "loss": 3.6889238357543945, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.017113137990236282, + "learning_rate": 0.0004493400204398426, + "loss": 3.684070587158203, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.017993126064538956, + "learning_rate": 0.00044910798221106283, + "loss": 3.675426721572876, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.018354708328843117, + "learning_rate": 0.00044887582626572105, + "loss": 3.675588607788086, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.016362832859158516, + "learning_rate": 0.0004486435527908606, + "loss": 3.6709303855895996, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01772182621061802, + "learning_rate": 0.0004484111619736205, + "loss": 3.6884102821350098, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.019271764904260635, + "learning_rate": 0.0004481786540012336, + "loss": 3.6468968391418457, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.019040528684854507, + "learning_rate": 0.00044794602906102747, + "loss": 3.690387010574341, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.019022610038518906, + "learning_rate": 0.00044771328734042367, + "loss": 3.689873218536377, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.018533391878008842, + "learning_rate": 0.00044748042902693815, + "loss": 3.6877284049987793, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.020825615152716637, + "learning_rate": 0.0004472474543081805, + "loss": 3.662257671356201, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.021880341693758965, + "learning_rate": 0.00044701436337185444, + "loss": 3.650700092315674, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.019974704831838608, + "learning_rate": 0.00044678115640575704, + "loss": 3.684177875518799, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.01903127320110798, + "learning_rate": 0.0004465478335977788, + "loss": 3.658148765563965, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.020999105647206306, + "learning_rate": 0.00044631439513590393, + "loss": 3.68747615814209, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.02138398587703705, + "learning_rate": 0.00044608084120820933, + "loss": 3.656355381011963, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.02216442860662937, + "learning_rate": 0.0004458471720028655, + "loss": 3.6860342025756836, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.019359121099114418, + "learning_rate": 0.0004456133877081352, + "loss": 3.654944658279419, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.020984770730137825, + "learning_rate": 0.0004453794885123744, + "loss": 3.6885194778442383, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.02423972077667713, + "learning_rate": 0.00044514547460403135, + "loss": 3.672712802886963, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.024544453248381615, + "learning_rate": 0.00044491134617164694, + "loss": 3.6484768390655518, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.060352325439453, + "eval_runtime": 57.9556, + "eval_samples_per_second": 42.136, + "eval_steps_per_second": 1.329, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.024072550237178802, + "learning_rate": 0.000444677103403854, + "loss": 3.5796468257904053, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.028701433911919594, + "learning_rate": 0.0004444427464893779, + "loss": 3.604318618774414, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.028119578957557678, + "learning_rate": 0.00044420827561703573, + "loss": 3.608759641647339, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.025335192680358887, + "learning_rate": 0.0004439736909757365, + "loss": 3.5760319232940674, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.0273929201066494, + "learning_rate": 0.0004437389927544805, + "loss": 3.559441089630127, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.026567714288830757, + "learning_rate": 0.0004435041811423603, + "loss": 3.626889228820801, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.027587885037064552, + "learning_rate": 0.0004432692563285591, + "loss": 3.6226272583007812, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.02935059368610382, + "learning_rate": 0.0004430342185023517, + "loss": 3.601921558380127, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.02982175536453724, + "learning_rate": 0.0004427990678531038, + "loss": 3.59737491607666, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.028570258989930153, + "learning_rate": 0.0004425638045702719, + "loss": 3.56337833404541, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.029890350997447968, + "learning_rate": 0.0004423284288434035, + "loss": 3.6294307708740234, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.03206814453005791, + "learning_rate": 0.00044209294086213657, + "loss": 3.590505599975586, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.03169345110654831, + "learning_rate": 0.00044185734081619957, + "loss": 3.581584930419922, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.03313758969306946, + "learning_rate": 0.000441621628895411, + "loss": 3.6088438034057617, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.03317603841423988, + "learning_rate": 0.00044138580528967986, + "loss": 3.6340456008911133, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.034738484770059586, + "learning_rate": 0.0004411498701890051, + "loss": 3.5814096927642822, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.036428093910217285, + "learning_rate": 0.0004409138237834751, + "loss": 3.5939526557922363, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.03293406963348389, + "learning_rate": 0.00044067766626326835, + "loss": 3.577897787094116, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.031030720099806786, + "learning_rate": 0.00044044139781865255, + "loss": 3.6077990531921387, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.03441236540675163, + "learning_rate": 0.00044020501863998516, + "loss": 3.591923713684082, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.03500965237617493, + "learning_rate": 0.00043996852891771236, + "loss": 3.6171927452087402, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.034443050622940063, + "learning_rate": 0.00043973192884236983, + "loss": 3.619724750518799, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.03435051813721657, + "learning_rate": 0.0004394952186045819, + "loss": 3.6443169116973877, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.03230702877044678, + "learning_rate": 0.00043925839839506167, + "loss": 3.5912671089172363, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.03232988342642784, + "learning_rate": 0.0004390214684046111, + "loss": 3.618985414505005, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.029298555105924606, + "learning_rate": 0.00043878442882412015, + "loss": 3.6166677474975586, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.026723450049757957, + "learning_rate": 0.00043854727984456746, + "loss": 3.61362886428833, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.028280874714255333, + "learning_rate": 0.0004383100216570197, + "loss": 3.6410951614379883, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.026700228452682495, + "learning_rate": 0.0004380726544526316, + "loss": 3.5804049968719482, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.023092713207006454, + "learning_rate": 0.00043783517842264534, + "loss": 3.595339298248291, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.024123897776007652, + "learning_rate": 0.0004375975937583914, + "loss": 3.647569417953491, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.02505572699010372, + "learning_rate": 0.0004373599006512871, + "loss": 3.6276652812957764, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.024523509666323662, + "learning_rate": 0.00043712209929283786, + "loss": 3.59525728225708, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.024351581931114197, + "learning_rate": 0.00043688418987463567, + "loss": 3.6081674098968506, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.02530614472925663, + "learning_rate": 0.00043664617258835995, + "loss": 3.6632542610168457, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.023438062518835068, + "learning_rate": 0.0004364080476257769, + "loss": 3.6345431804656982, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.021853657439351082, + "learning_rate": 0.0004361698151787396, + "loss": 3.598644256591797, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.022487422451376915, + "learning_rate": 0.0004359314754391874, + "loss": 3.5966577529907227, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.021353935822844505, + "learning_rate": 0.0004356930285991466, + "loss": 3.6328375339508057, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.021180452778935432, + "learning_rate": 0.00043545447485072926, + "loss": 3.6285812854766846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.020491832867264748, + "learning_rate": 0.000435215814386134, + "loss": 3.592944622039795, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.021603770554065704, + "learning_rate": 0.0004349770473976453, + "loss": 3.613507032394409, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.02146158367395401, + "learning_rate": 0.0004347381740776332, + "loss": 3.6409506797790527, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.02102433145046234, + "learning_rate": 0.0004344991946185538, + "loss": 3.6179330348968506, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.018684551119804382, + "learning_rate": 0.00043426010921294837, + "loss": 3.6385130882263184, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.018416227772831917, + "learning_rate": 0.0004340209180534438, + "loss": 3.642151355743408, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.01915144734084606, + "learning_rate": 0.000433781621332752, + "loss": 3.6089656352996826, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.018299665302038193, + "learning_rate": 0.0004335422192436703, + "loss": 3.6063547134399414, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.018962757661938667, + "learning_rate": 0.00043330271197908023, + "loss": 3.625917434692383, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.017437821254134178, + "learning_rate": 0.00043306309973194874, + "loss": 3.600015640258789, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.017913147807121277, + "learning_rate": 0.000432823382695327, + "loss": 3.6428744792938232, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.018343618139624596, + "learning_rate": 0.0004325835610623508, + "loss": 3.635901927947998, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.020372850820422173, + "learning_rate": 0.00043234363502623994, + "loss": 3.620181083679199, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.02037600241601467, + "learning_rate": 0.0004321036047802985, + "loss": 3.61129093170166, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.021827151998877525, + "learning_rate": 0.00043186347051791466, + "loss": 3.5999555587768555, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.021806657314300537, + "learning_rate": 0.0004316232324325602, + "loss": 3.642509937286377, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02050699107348919, + "learning_rate": 0.0004313828907177906, + "loss": 3.597583293914795, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.020436758175492287, + "learning_rate": 0.0004311424455672448, + "loss": 3.6079912185668945, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.019341666251420975, + "learning_rate": 0.0004309018971746451, + "loss": 3.6196017265319824, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.022348817437887192, + "learning_rate": 0.000430661245733797, + "loss": 3.619025468826294, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.019361121580004692, + "learning_rate": 0.00043042049143858914, + "loss": 3.657182455062866, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.019258935004472733, + "learning_rate": 0.0004301796344829928, + "loss": 3.6408963203430176, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.020622272044420242, + "learning_rate": 0.00042993867506106205, + "loss": 3.6247220039367676, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.019510755315423012, + "learning_rate": 0.0004296976133669336, + "loss": 3.6264584064483643, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.02078665979206562, + "learning_rate": 0.0004294564495948266, + "loss": 3.6364388465881348, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01791800744831562, + "learning_rate": 0.0004292151839390421, + "loss": 3.6029374599456787, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.01922667771577835, + "learning_rate": 0.0004289738165939636, + "loss": 3.582937717437744, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.018473893404006958, + "learning_rate": 0.00042873234775405654, + "loss": 3.6141819953918457, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01846298575401306, + "learning_rate": 0.00042849077761386784, + "loss": 3.638786554336548, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.01876973733305931, + "learning_rate": 0.0004282491063680263, + "loss": 3.617130756378174, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.017995499074459076, + "learning_rate": 0.00042800733421124207, + "loss": 3.645305633544922, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.02105429396033287, + "learning_rate": 0.00042776546133830646, + "loss": 3.6444876194000244, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.049062252044678, + "eval_runtime": 58.0113, + "eval_samples_per_second": 42.095, + "eval_steps_per_second": 1.327, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.02128119394183159, + "learning_rate": 0.00042752348794409227, + "loss": 3.524954080581665, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.028249500319361687, + "learning_rate": 0.00042728141422355287, + "loss": 3.5311479568481445, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.03145391866564751, + "learning_rate": 0.0004270392403717229, + "loss": 3.559330940246582, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.02994118444621563, + "learning_rate": 0.00042679696658371743, + "loss": 3.5257620811462402, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.027448246255517006, + "learning_rate": 0.00042655459305473195, + "loss": 3.5242764949798584, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.0296134352684021, + "learning_rate": 0.0004263121199800425, + "loss": 3.5317091941833496, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.029576804488897324, + "learning_rate": 0.0004260695475550054, + "loss": 3.5340428352355957, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.02941269613802433, + "learning_rate": 0.0004258268759750566, + "loss": 3.5372023582458496, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.03126693144440651, + "learning_rate": 0.0004255841054357124, + "loss": 3.5437135696411133, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.03335367515683174, + "learning_rate": 0.0004253412361325686, + "loss": 3.5598273277282715, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.03028946928679943, + "learning_rate": 0.0004250982682613006, + "loss": 3.5190839767456055, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.028629133477807045, + "learning_rate": 0.00042485520201766315, + "loss": 3.5138821601867676, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.027070026844739914, + "learning_rate": 0.0004246120375974905, + "loss": 3.530958414077759, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.028837760910391808, + "learning_rate": 0.0004243687751966957, + "loss": 3.5233962535858154, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.029884297400712967, + "learning_rate": 0.0004241254150112708, + "loss": 3.5639758110046387, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.024223247542977333, + "learning_rate": 0.000423881957237287, + "loss": 3.5555152893066406, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.026992836967110634, + "learning_rate": 0.0004236384020708935, + "loss": 3.540196418762207, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.02818330004811287, + "learning_rate": 0.0004233947497083185, + "loss": 3.5546071529388428, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.026401042938232422, + "learning_rate": 0.0004231510003458682, + "loss": 3.5556247234344482, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.026020489633083344, + "learning_rate": 0.0004229071541799272, + "loss": 3.568004608154297, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.02803782932460308, + "learning_rate": 0.0004226632114069578, + "loss": 3.5828235149383545, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.02737978659570217, + "learning_rate": 0.0004224191722235004, + "loss": 3.5642499923706055, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.02406371757388115, + "learning_rate": 0.00042217503682617286, + "loss": 3.5530993938446045, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.02230692468583584, + "learning_rate": 0.0004219308054116706, + "loss": 3.5573501586914062, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.021352406591176987, + "learning_rate": 0.0004216864781767667, + "loss": 3.5526974201202393, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.02227999083697796, + "learning_rate": 0.00042144205531831103, + "loss": 3.5866928100585938, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.024190247058868408, + "learning_rate": 0.0004211975370332308, + "loss": 3.517117977142334, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.023196619004011154, + "learning_rate": 0.00042095292351852976, + "loss": 3.4826502799987793, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.02388714998960495, + "learning_rate": 0.0004207082149712888, + "loss": 3.5588626861572266, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.024692339822649956, + "learning_rate": 0.0004204634115886651, + "loss": 3.54828143119812, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.023078206926584244, + "learning_rate": 0.0004202185135678924, + "loss": 3.5679030418395996, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.020665735006332397, + "learning_rate": 0.00041997352110628033, + "loss": 3.5500283241271973, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.02022322081029415, + "learning_rate": 0.00041972843440121527, + "loss": 3.592104911804199, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.021137656643986702, + "learning_rate": 0.00041948325365015894, + "loss": 3.583833694458008, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.021026931703090668, + "learning_rate": 0.0004192379790506491, + "loss": 3.556403875350952, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.020423648878932, + "learning_rate": 0.00041899261080029904, + "loss": 3.5839173793792725, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.020935002714395523, + "learning_rate": 0.00041874714909679754, + "loss": 3.5794217586517334, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.020799759775400162, + "learning_rate": 0.00041850159413790863, + "loss": 3.6033859252929688, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.019480884075164795, + "learning_rate": 0.0004182559461214715, + "loss": 3.580097198486328, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.019300777465105057, + "learning_rate": 0.0004180102052454004, + "loss": 3.550474166870117, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.020265772938728333, + "learning_rate": 0.00041776437170768427, + "loss": 3.5711655616760254, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.019972780719399452, + "learning_rate": 0.0004175184457063868, + "loss": 3.5636885166168213, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.019635623320937157, + "learning_rate": 0.000417272427439646, + "loss": 3.5851311683654785, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.021451901644468307, + "learning_rate": 0.00041702631710567443, + "loss": 3.5956759452819824, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.019456414505839348, + "learning_rate": 0.00041678011490275875, + "loss": 3.595731258392334, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.019757689908146858, + "learning_rate": 0.00041653382102925957, + "loss": 3.569652557373047, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.020571604371070862, + "learning_rate": 0.00041628743568361147, + "loss": 3.6007747650146484, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.02153848111629486, + "learning_rate": 0.00041604095906432265, + "loss": 3.5915260314941406, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.020875850692391396, + "learning_rate": 0.00041579439136997493, + "loss": 3.5582690238952637, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.02089419960975647, + "learning_rate": 0.00041554773279922337, + "loss": 3.5756618976593018, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01961098238825798, + "learning_rate": 0.0004153009835507963, + "loss": 3.5918776988983154, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.021075716242194176, + "learning_rate": 0.0004150541438234952, + "loss": 3.5342202186584473, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.020745476707816124, + "learning_rate": 0.00041480721381619435, + "loss": 3.5976028442382812, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.02001352794468403, + "learning_rate": 0.00041456019372784086, + "loss": 3.5686182975769043, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.021684160456061363, + "learning_rate": 0.0004143130837574543, + "loss": 3.553445339202881, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.022551920264959335, + "learning_rate": 0.0004140658841041267, + "loss": 3.5712709426879883, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.022758224979043007, + "learning_rate": 0.0004138185949670224, + "loss": 3.6101667881011963, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.020714636892080307, + "learning_rate": 0.00041357121654537766, + "loss": 3.5740506649017334, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.022694187238812447, + "learning_rate": 0.00041332374903850104, + "loss": 3.5663979053497314, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.021664340049028397, + "learning_rate": 0.00041307619264577234, + "loss": 3.5830888748168945, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.019072270020842552, + "learning_rate": 0.0004128285475666436, + "loss": 3.5861973762512207, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.020153764635324478, + "learning_rate": 0.0004125808140006378, + "loss": 3.566444158554077, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019666558131575584, + "learning_rate": 0.0004123329921473494, + "loss": 3.5789592266082764, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.02101624570786953, + "learning_rate": 0.0004120850822064439, + "loss": 3.551722526550293, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.01951020397245884, + "learning_rate": 0.00041183708437765815, + "loss": 3.6069014072418213, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.01985476352274418, + "learning_rate": 0.00041158899886079925, + "loss": 3.5944924354553223, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02195335365831852, + "learning_rate": 0.00041134082585574534, + "loss": 3.5828046798706055, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02105172537267208, + "learning_rate": 0.00041109256556244497, + "loss": 3.5830953121185303, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.01888827420771122, + "learning_rate": 0.00041084421818091685, + "loss": 3.5656847953796387, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.017867228016257286, + "learning_rate": 0.0004105957839112501, + "loss": 3.5792243480682373, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.018577905371785164, + "learning_rate": 0.0004103472629536036, + "loss": 3.56992506980896, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.020045632496476173, + "learning_rate": 0.0004100986555082063, + "loss": 3.590876579284668, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.044339656829834, + "eval_runtime": 57.1435, + "eval_samples_per_second": 42.735, + "eval_steps_per_second": 1.347, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.024135645478963852, + "learning_rate": 0.0004098499617753567, + "loss": 3.4856467247009277, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.031053155660629272, + "learning_rate": 0.00040960118195542273, + "loss": 3.455773115158081, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.035190410912036896, + "learning_rate": 0.00040935231624884204, + "loss": 3.506619930267334, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.03611094877123833, + "learning_rate": 0.0004091033648561211, + "loss": 3.5227696895599365, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.034121073782444, + "learning_rate": 0.00040885432797783543, + "loss": 3.4970145225524902, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.03629382699728012, + "learning_rate": 0.0004086052058146297, + "loss": 3.506425619125366, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.03709045797586441, + "learning_rate": 0.00040835599856721725, + "loss": 3.499274969100952, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.034826792776584625, + "learning_rate": 0.00040810670643637954, + "loss": 3.5029163360595703, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.03344690054655075, + "learning_rate": 0.000407857329622967, + "loss": 3.4816911220550537, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.03136949986219406, + "learning_rate": 0.00040760786832789786, + "loss": 3.5154659748077393, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.030533097684383392, + "learning_rate": 0.0004073583227521588, + "loss": 3.509666919708252, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.03298318013548851, + "learning_rate": 0.0004071086930968039, + "loss": 3.5096616744995117, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.031186232343316078, + "learning_rate": 0.00040685897956295545, + "loss": 3.4752590656280518, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.028008371591567993, + "learning_rate": 0.0004066091823518031, + "loss": 3.498826503753662, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.028206519782543182, + "learning_rate": 0.00040635930166460385, + "loss": 3.5281882286071777, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.024417119100689888, + "learning_rate": 0.00040610933770268226, + "loss": 3.506791830062866, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.025321070104837418, + "learning_rate": 0.00040585929066742964, + "loss": 3.494256019592285, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.023430505767464638, + "learning_rate": 0.00040560916076030435, + "loss": 3.533010482788086, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.023371374234557152, + "learning_rate": 0.00040535894818283154, + "loss": 3.5197086334228516, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.02414027415215969, + "learning_rate": 0.0004051086531366031, + "loss": 3.4707889556884766, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.026652060449123383, + "learning_rate": 0.000404858275823277, + "loss": 3.498159885406494, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.024486590176820755, + "learning_rate": 0.000404607816444578, + "loss": 3.50034236907959, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.023153992369771004, + "learning_rate": 0.00040435727520229644, + "loss": 3.499504804611206, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.02222427912056446, + "learning_rate": 0.0004041066522982891, + "loss": 3.4953579902648926, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.023033641278743744, + "learning_rate": 0.000403855947934478, + "loss": 3.5181496143341064, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.02141181193292141, + "learning_rate": 0.00040360516231285155, + "loss": 3.552860736846924, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.023448696359992027, + "learning_rate": 0.00040335429563546286, + "loss": 3.5153324604034424, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.022410666570067406, + "learning_rate": 0.0004031033481044308, + "loss": 3.4963512420654297, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.020326802507042885, + "learning_rate": 0.00040285231992193924, + "loss": 3.528319835662842, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021117394790053368, + "learning_rate": 0.0004026012112902372, + "loss": 3.5192837715148926, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.021317072212696075, + "learning_rate": 0.0004023500224116381, + "loss": 3.5081021785736084, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.021116144955158234, + "learning_rate": 0.00040209875348852037, + "loss": 3.5167174339294434, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019834019243717194, + "learning_rate": 0.00040184740472332705, + "loss": 3.511960983276367, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.02089378982782364, + "learning_rate": 0.000401595976318565, + "loss": 3.493330955505371, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.021129710599780083, + "learning_rate": 0.00040134446847680584, + "loss": 3.544156551361084, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.021163299679756165, + "learning_rate": 0.0004010928814006846, + "loss": 3.5261003971099854, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.02105306088924408, + "learning_rate": 0.0004008412152929007, + "loss": 3.543973445892334, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.02150859497487545, + "learning_rate": 0.0004005894703562168, + "loss": 3.487217426300049, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.021666452288627625, + "learning_rate": 0.0004003376467934593, + "loss": 3.4933671951293945, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.02180723287165165, + "learning_rate": 0.0004000857448075178, + "loss": 3.55198073387146, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.02041625790297985, + "learning_rate": 0.0003998337646013455, + "loss": 3.529571533203125, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.022977063432335854, + "learning_rate": 0.00039958170637795806, + "loss": 3.526432752609253, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.022725161164999008, + "learning_rate": 0.0003993295703404341, + "loss": 3.5258829593658447, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.021337240934371948, + "learning_rate": 0.00039907735669191544, + "loss": 3.4897351264953613, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.020783131942152977, + "learning_rate": 0.00039882506563560573, + "loss": 3.5497589111328125, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.022259317338466644, + "learning_rate": 0.0003985726973747715, + "loss": 3.5225064754486084, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.020283272489905357, + "learning_rate": 0.00039832025211274113, + "loss": 3.5453078746795654, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.020656265318393707, + "learning_rate": 0.00039806773005290544, + "loss": 3.533703327178955, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.02104976586997509, + "learning_rate": 0.0003978151313987168, + "loss": 3.4885146617889404, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.021411150693893433, + "learning_rate": 0.00039756245635368933, + "loss": 3.533268690109253, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.021906673908233643, + "learning_rate": 0.00039730970512139876, + "loss": 3.5483224391937256, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.02415728010237217, + "learning_rate": 0.0003970568779054822, + "loss": 3.5084290504455566, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.023435061797499657, + "learning_rate": 0.000396803974909638, + "loss": 3.5498428344726562, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.02183385379612446, + "learning_rate": 0.0003965509963376255, + "loss": 3.484086751937866, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.022856134921312332, + "learning_rate": 0.000396297942393265, + "loss": 3.5339231491088867, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.021679896861314774, + "learning_rate": 0.0003960448132804375, + "loss": 3.547191619873047, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.021262124180793762, + "learning_rate": 0.0003957916092030845, + "loss": 3.5263822078704834, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.02265402302145958, + "learning_rate": 0.00039553833036520803, + "loss": 3.5361135005950928, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.020080721005797386, + "learning_rate": 0.0003952849769708702, + "loss": 3.5247459411621094, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.02187187783420086, + "learning_rate": 0.0003950315492241932, + "loss": 3.54129695892334, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.02417500875890255, + "learning_rate": 0.0003947780473293593, + "loss": 3.598796844482422, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.02416270412504673, + "learning_rate": 0.00039452447149061054, + "loss": 3.5207936763763428, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.021958643570542336, + "learning_rate": 0.0003942708219122481, + "loss": 3.5395452976226807, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.021813420578837395, + "learning_rate": 0.000394017098798633, + "loss": 3.5469255447387695, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020607495680451393, + "learning_rate": 0.0003937633023541854, + "loss": 3.554851531982422, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.02091527171432972, + "learning_rate": 0.0003935094327833845, + "loss": 3.492576837539673, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.021089257672429085, + "learning_rate": 0.0003932554902907683, + "loss": 3.5154709815979004, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.019659532234072685, + "learning_rate": 0.0003930014750809338, + "loss": 3.5444283485412598, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01896163448691368, + "learning_rate": 0.0003927473873585365, + "loss": 3.589625358581543, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020288823172450066, + "learning_rate": 0.0003924932273282903, + "loss": 3.522411823272705, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.020222362130880356, + "learning_rate": 0.00039223899519496723, + "loss": 3.5244359970092773, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.022349944338202477, + "learning_rate": 0.00039198469116339757, + "loss": 3.5362396240234375, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.047028064727783, + "eval_runtime": 62.8948, + "eval_samples_per_second": 38.827, + "eval_steps_per_second": 1.224, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.02202913723886013, + "learning_rate": 0.0003917303154384694, + "loss": 3.4249138832092285, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.03153969347476959, + "learning_rate": 0.00039147586822512885, + "loss": 3.4436073303222656, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.03747870773077011, + "learning_rate": 0.0003912213497283793, + "loss": 3.4879579544067383, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.036144088953733444, + "learning_rate": 0.0003909667601532819, + "loss": 3.419151782989502, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.03983336314558983, + "learning_rate": 0.00039071209970495445, + "loss": 3.4331436157226562, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.045912645757198334, + "learning_rate": 0.00039045736858857273, + "loss": 3.4304769039154053, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.04005753993988037, + "learning_rate": 0.0003902025670093687, + "loss": 3.478600025177002, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.03488151356577873, + "learning_rate": 0.00038994769517263156, + "loss": 3.4650232791900635, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.03603382781147957, + "learning_rate": 0.0003896927532837069, + "loss": 3.449807643890381, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.03817906603217125, + "learning_rate": 0.0003894377415479967, + "loss": 3.4556589126586914, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.03523671627044678, + "learning_rate": 0.00038918266017095936, + "loss": 3.477426767349243, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.03284068778157234, + "learning_rate": 0.0003889275093581096, + "loss": 3.478672981262207, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.02992352657020092, + "learning_rate": 0.00038867228931501737, + "loss": 3.4810047149658203, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.02858242578804493, + "learning_rate": 0.0003884170002473091, + "loss": 3.467811107635498, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.027912171557545662, + "learning_rate": 0.0003881616423606666, + "loss": 3.4326171875, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.028351375833153725, + "learning_rate": 0.00038790621586082697, + "loss": 3.4533729553222656, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.027947887778282166, + "learning_rate": 0.0003876507209535829, + "loss": 3.420295476913452, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.028510035946965218, + "learning_rate": 0.00038739515784478186, + "loss": 3.438023328781128, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.027707934379577637, + "learning_rate": 0.00038713952674032654, + "loss": 3.4844181537628174, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.0262764822691679, + "learning_rate": 0.0003868838278461743, + "loss": 3.4363749027252197, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.02713681012392044, + "learning_rate": 0.0003866280613683371, + "loss": 3.4501900672912598, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.0299485195428133, + "learning_rate": 0.00038637222751288126, + "loss": 3.461942672729492, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.027047986164689064, + "learning_rate": 0.0003861163264859277, + "loss": 3.454293727874756, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.02436736971139908, + "learning_rate": 0.0003858603584936511, + "loss": 3.4663405418395996, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.02601189725100994, + "learning_rate": 0.0003856043237422803, + "loss": 3.502030849456787, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.02463468722999096, + "learning_rate": 0.0003853482224380978, + "loss": 3.462491035461426, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.021793190389871597, + "learning_rate": 0.00038509205478744, + "loss": 3.4476377964019775, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.023809107020497322, + "learning_rate": 0.0003848358209966963, + "loss": 3.4701178073883057, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.02557182125747204, + "learning_rate": 0.0003845795212723099, + "loss": 3.4932539463043213, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.023424552753567696, + "learning_rate": 0.00038432315582077664, + "loss": 3.4779105186462402, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.02195148915052414, + "learning_rate": 0.0003840667248486456, + "loss": 3.470506191253662, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.022897250950336456, + "learning_rate": 0.0003838102285625186, + "loss": 3.48982572555542, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.023033805191516876, + "learning_rate": 0.00038355366716905006, + "loss": 3.4845101833343506, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.022774485871195793, + "learning_rate": 0.0003832970408749467, + "loss": 3.5060324668884277, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.02193611115217209, + "learning_rate": 0.00038304034988696816, + "loss": 3.511518955230713, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.02130570448935032, + "learning_rate": 0.00038278359441192516, + "loss": 3.468502998352051, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.020402755588293076, + "learning_rate": 0.00038252677465668136, + "loss": 3.4786908626556396, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.022664198651909828, + "learning_rate": 0.00038226989082815156, + "loss": 3.479928970336914, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.02454056777060032, + "learning_rate": 0.00038201294313330245, + "loss": 3.479947566986084, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.022096732631325722, + "learning_rate": 0.00038175593177915227, + "loss": 3.49478816986084, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.021499406546354294, + "learning_rate": 0.0003814988569727704, + "loss": 3.494933605194092, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.022325703874230385, + "learning_rate": 0.0003812417189212773, + "loss": 3.5041463375091553, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.020674455910921097, + "learning_rate": 0.00038098451783184447, + "loss": 3.472404718399048, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.02131054364144802, + "learning_rate": 0.00038072725391169416, + "loss": 3.497795343399048, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.02126218006014824, + "learning_rate": 0.0003804699273680994, + "loss": 3.483281373977661, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.0217830128967762, + "learning_rate": 0.0003802125384083834, + "loss": 3.4704298973083496, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.02156917378306389, + "learning_rate": 0.0003799550872399197, + "loss": 3.4994399547576904, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.02047055773437023, + "learning_rate": 0.0003796975740701323, + "loss": 3.4316084384918213, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.019766805693507195, + "learning_rate": 0.0003794399991064948, + "loss": 3.452223777770996, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.020028885453939438, + "learning_rate": 0.00037918236255653074, + "loss": 3.485278844833374, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.020312055945396423, + "learning_rate": 0.00037892466462781306, + "loss": 3.4789156913757324, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.020725280046463013, + "learning_rate": 0.00037866690552796446, + "loss": 3.495528221130371, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.02088077738881111, + "learning_rate": 0.00037840908546465696, + "loss": 3.4972569942474365, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.021010037511587143, + "learning_rate": 0.0003781512046456111, + "loss": 3.473562240600586, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.019641095772385597, + "learning_rate": 0.0003778932632785972, + "loss": 3.4782915115356445, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.02063329517841339, + "learning_rate": 0.00037763526157143376, + "loss": 3.4681782722473145, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.0213928259909153, + "learning_rate": 0.00037737719973198825, + "loss": 3.472263813018799, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.02135022170841694, + "learning_rate": 0.0003771190779681762, + "loss": 3.502300262451172, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.02053583413362503, + "learning_rate": 0.0003768608964879619, + "loss": 3.489243745803833, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.021575238555669785, + "learning_rate": 0.00037660265549935724, + "loss": 3.491074562072754, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.022335771471261978, + "learning_rate": 0.00037634435521042257, + "loss": 3.5298843383789062, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.021198464557528496, + "learning_rate": 0.0003760859958292656, + "loss": 3.5156898498535156, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.020986037328839302, + "learning_rate": 0.000375827577564042, + "loss": 3.4574475288391113, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.022639811038970947, + "learning_rate": 0.0003755691006229545, + "loss": 3.4952611923217773, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.019711066037416458, + "learning_rate": 0.0003753105652142534, + "loss": 3.531296730041504, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.02058061584830284, + "learning_rate": 0.000375051971546236, + "loss": 3.5340657234191895, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.019849255681037903, + "learning_rate": 0.0003747933198272465, + "loss": 3.5028295516967773, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01954430714249611, + "learning_rate": 0.00037453461026567604, + "loss": 3.51088809967041, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.019234873354434967, + "learning_rate": 0.00037427584306996196, + "loss": 3.472560167312622, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.021085912361741066, + "learning_rate": 0.0003740170184485888, + "loss": 3.468876838684082, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.020300406962633133, + "learning_rate": 0.0003737581366100864, + "loss": 3.4903886318206787, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02380351349711418, + "learning_rate": 0.0003734991977630315, + "loss": 3.5069704055786133, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.048111915588379, + "eval_runtime": 58.312, + "eval_samples_per_second": 41.878, + "eval_steps_per_second": 1.32, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02379327267408371, + "learning_rate": 0.0003732402021160463, + "loss": 3.396376132965088, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.03204222396016121, + "learning_rate": 0.00037298114987779885, + "loss": 3.4102582931518555, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.034993767738342285, + "learning_rate": 0.000372722041257003, + "loss": 3.385697364807129, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.03387003391981125, + "learning_rate": 0.00037246287646241783, + "loss": 3.4120078086853027, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.03282478079199791, + "learning_rate": 0.0003722036557028478, + "loss": 3.404122829437256, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.03053346462547779, + "learning_rate": 0.0003719443791871422, + "loss": 3.368051767349243, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.029722390696406364, + "learning_rate": 0.0003716850471241958, + "loss": 3.4066247940063477, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.0287226140499115, + "learning_rate": 0.0003714256597229474, + "loss": 3.4212646484375, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.028156550601124763, + "learning_rate": 0.0003711662171923809, + "loss": 3.4192593097686768, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.03021939843893051, + "learning_rate": 0.0003709067197415244, + "loss": 3.3973188400268555, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.03314025700092316, + "learning_rate": 0.00037064716757945036, + "loss": 3.422978401184082, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.03156686946749687, + "learning_rate": 0.0003703875609152753, + "loss": 3.4052343368530273, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.03357478231191635, + "learning_rate": 0.0003701278999581595, + "loss": 3.383730411529541, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.03261081501841545, + "learning_rate": 0.0003698681849173073, + "loss": 3.421250820159912, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.03252755105495453, + "learning_rate": 0.0003696084160019662, + "loss": 3.39884352684021, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.03177274018526077, + "learning_rate": 0.0003693485934214274, + "loss": 3.4175422191619873, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.028845487162470818, + "learning_rate": 0.0003690887173850253, + "loss": 3.3908143043518066, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.026045162230730057, + "learning_rate": 0.0003688287881021374, + "loss": 3.4417710304260254, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.027937524020671844, + "learning_rate": 0.00036856880578218376, + "loss": 3.410548210144043, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.025780271738767624, + "learning_rate": 0.00036830877063462786, + "loss": 3.4069018363952637, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.026093080639839172, + "learning_rate": 0.0003680486828689749, + "loss": 3.4199600219726562, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.025765180587768555, + "learning_rate": 0.00036778854269477315, + "loss": 3.442903995513916, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.022773101925849915, + "learning_rate": 0.0003675283503216127, + "loss": 3.394232749938965, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.025135623291134834, + "learning_rate": 0.000367268105959126, + "loss": 3.4466171264648438, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.02491890639066696, + "learning_rate": 0.00036700780981698705, + "loss": 3.418689727783203, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.02521841786801815, + "learning_rate": 0.0003667474621049119, + "loss": 3.4213061332702637, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.025186199694871902, + "learning_rate": 0.00036648706303265795, + "loss": 3.437631130218506, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.022097431123256683, + "learning_rate": 0.000366226612810024, + "loss": 3.4014902114868164, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.024009618908166885, + "learning_rate": 0.00036596611164685025, + "loss": 3.4292187690734863, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.022758696228265762, + "learning_rate": 0.0003657055597530175, + "loss": 3.4105381965637207, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.02447321265935898, + "learning_rate": 0.00036544495733844797, + "loss": 3.4196958541870117, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.025843804702162743, + "learning_rate": 0.000365184304613104, + "loss": 3.4343783855438232, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.022433951497077942, + "learning_rate": 0.0003649236017869892, + "loss": 3.440580368041992, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.02250702679157257, + "learning_rate": 0.00036466284907014685, + "loss": 3.4255261421203613, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.022953296080231667, + "learning_rate": 0.0003644020466726608, + "loss": 3.442258358001709, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.021977661177515984, + "learning_rate": 0.00036414119480465473, + "loss": 3.4058687686920166, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.021751178428530693, + "learning_rate": 0.0003638802936762925, + "loss": 3.4211649894714355, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.022764205932617188, + "learning_rate": 0.0003636193434977771, + "loss": 3.424694061279297, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.024827787652611732, + "learning_rate": 0.0003633583444793516, + "loss": 3.458076238632202, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.021888602524995804, + "learning_rate": 0.00036309729683129814, + "loss": 3.4509153366088867, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.021961702033877373, + "learning_rate": 0.00036283620076393814, + "loss": 3.4147253036499023, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.02235778421163559, + "learning_rate": 0.000362575056487632, + "loss": 3.4392640590667725, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.021714260801672935, + "learning_rate": 0.00036231386421277877, + "loss": 3.444026470184326, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.020996905863285065, + "learning_rate": 0.0003620526241498167, + "loss": 3.4168434143066406, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.020528919994831085, + "learning_rate": 0.0003617913365092218, + "loss": 3.4831130504608154, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.021056141704320908, + "learning_rate": 0.0003615300015015091, + "loss": 3.430905342102051, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.021655110642313957, + "learning_rate": 0.0003612686193372312, + "loss": 3.4588241577148438, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.02344612590968609, + "learning_rate": 0.00036100719022697924, + "loss": 3.4202680587768555, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.02264723740518093, + "learning_rate": 0.0003607457143813819, + "loss": 3.4669015407562256, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.02377854660153389, + "learning_rate": 0.00036048419201110553, + "loss": 3.5059216022491455, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.02299923449754715, + "learning_rate": 0.00036022262332685383, + "loss": 3.4796524047851562, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.021110547706484795, + "learning_rate": 0.0003599610085393681, + "loss": 3.4342408180236816, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.022060122340917587, + "learning_rate": 0.0003596993478594267, + "loss": 3.426923990249634, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.021077830344438553, + "learning_rate": 0.0003594376414978447, + "loss": 3.4636270999908447, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.02209082990884781, + "learning_rate": 0.0003591758896654745, + "loss": 3.441979169845581, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.022408071905374527, + "learning_rate": 0.0003589140925732045, + "loss": 3.490835666656494, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.022462666034698486, + "learning_rate": 0.0003586522504319602, + "loss": 3.452770233154297, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.021295800805091858, + "learning_rate": 0.0003583903634527029, + "loss": 3.4377408027648926, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.020457757636904716, + "learning_rate": 0.0003581284318464302, + "loss": 3.4191250801086426, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.021163633093237877, + "learning_rate": 0.00035786645582417564, + "loss": 3.4797091484069824, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.021495461463928223, + "learning_rate": 0.00035760443559700863, + "loss": 3.4763145446777344, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.021050071343779564, + "learning_rate": 0.00035734237137603417, + "loss": 3.4687209129333496, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.02150929905474186, + "learning_rate": 0.0003570802633723927, + "loss": 3.4819254875183105, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.0212967898696661, + "learning_rate": 0.0003568181117972597, + "loss": 3.490407705307007, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.02123691327869892, + "learning_rate": 0.00035655591686184626, + "loss": 3.4418535232543945, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.02193271741271019, + "learning_rate": 0.0003562936787773979, + "loss": 3.446892738342285, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.02065582573413849, + "learning_rate": 0.000356031397755195, + "loss": 3.4308691024780273, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.019487040117383003, + "learning_rate": 0.00035576907400655303, + "loss": 3.487239122390747, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.021190891042351723, + "learning_rate": 0.00035550670774282106, + "loss": 3.4635558128356934, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.0209872554987669, + "learning_rate": 0.0003552442991753831, + "loss": 3.443793773651123, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.02035645768046379, + "learning_rate": 0.00035498184851565697, + "loss": 3.473198652267456, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.024002201855182648, + "learning_rate": 0.0003547193559750944, + "loss": 3.4244751930236816, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.057085990905762, + "eval_runtime": 62.4735, + "eval_samples_per_second": 39.089, + "eval_steps_per_second": 1.233, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.02450242079794407, + "learning_rate": 0.00035445682176518087, + "loss": 3.361237049102783, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.030813416466116905, + "learning_rate": 0.00035419424609743546, + "loss": 3.365708351135254, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.03279578685760498, + "learning_rate": 0.00035393162918341065, + "loss": 3.325385570526123, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.03293970227241516, + "learning_rate": 0.00035366897123469213, + "loss": 3.356811285018921, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.033190157264471054, + "learning_rate": 0.0003534062724628986, + "loss": 3.406217098236084, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.033370476216077805, + "learning_rate": 0.00035314353307968195, + "loss": 3.3579213619232178, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.03351330757141113, + "learning_rate": 0.00035288075329672634, + "loss": 3.3755106925964355, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.034596871584653854, + "learning_rate": 0.0003526179333257487, + "loss": 3.3576536178588867, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.03177076578140259, + "learning_rate": 0.0003523550733784984, + "loss": 3.3123257160186768, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.0289607010781765, + "learning_rate": 0.00035209217366675706, + "loss": 3.370959997177124, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.03265267238020897, + "learning_rate": 0.000351829234402338, + "loss": 3.332608938217163, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.03397805988788605, + "learning_rate": 0.0003515662557970867, + "loss": 3.364213228225708, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.03100395016372204, + "learning_rate": 0.0003513032380628804, + "loss": 3.388005256652832, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.03224753588438034, + "learning_rate": 0.0003510401814116277, + "loss": 3.3267595767974854, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.03034576214849949, + "learning_rate": 0.0003507770860552684, + "loss": 3.3954129219055176, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.026316817849874496, + "learning_rate": 0.00035051395220577397, + "loss": 3.3543832302093506, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.027714375406503677, + "learning_rate": 0.0003502507800751464, + "loss": 3.4190099239349365, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.028731998056173325, + "learning_rate": 0.0003499875698754187, + "loss": 3.342031955718994, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02845810353755951, + "learning_rate": 0.00034972432181865467, + "loss": 3.356130599975586, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.02842753194272518, + "learning_rate": 0.00034946103611694854, + "loss": 3.3709869384765625, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.02942183054983616, + "learning_rate": 0.00034919771298242477, + "loss": 3.373588800430298, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.02958574891090393, + "learning_rate": 0.00034893435262723814, + "loss": 3.377514600753784, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.029524607583880424, + "learning_rate": 0.00034867095526357325, + "loss": 3.398663282394409, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.028092598542571068, + "learning_rate": 0.0003484075211036446, + "loss": 3.3742518424987793, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.026652829721570015, + "learning_rate": 0.0003481440503596964, + "loss": 3.387667179107666, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.024675756692886353, + "learning_rate": 0.0003478805432440021, + "loss": 3.3737070560455322, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.02491692826151848, + "learning_rate": 0.0003476169999688648, + "loss": 3.359217643737793, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.024601681157946587, + "learning_rate": 0.00034735342074661654, + "loss": 3.3902688026428223, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.02313913404941559, + "learning_rate": 0.0003470898057896183, + "loss": 3.3700339794158936, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.02418898604810238, + "learning_rate": 0.0003468261553102599, + "loss": 3.4265692234039307, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.023382030427455902, + "learning_rate": 0.00034656246952095984, + "loss": 3.378389835357666, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.02340470254421234, + "learning_rate": 0.0003462987486341648, + "loss": 3.392937421798706, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.022166339680552483, + "learning_rate": 0.00034603499286235006, + "loss": 3.39414119720459, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.022175129503011703, + "learning_rate": 0.0003457712024180188, + "loss": 3.407447099685669, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.02368207648396492, + "learning_rate": 0.0003455073775137025, + "loss": 3.415076971054077, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.02263982594013214, + "learning_rate": 0.0003452435183619598, + "loss": 3.383986473083496, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.02147712931036949, + "learning_rate": 0.0003449796251753773, + "loss": 3.403963565826416, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.022987166419625282, + "learning_rate": 0.00034471569816656915, + "loss": 3.3883113861083984, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.02240179106593132, + "learning_rate": 0.00034445173754817646, + "loss": 3.4349911212921143, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.02206592820584774, + "learning_rate": 0.00034418774353286747, + "loss": 3.426473617553711, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.022650036960840225, + "learning_rate": 0.0003439237163333375, + "loss": 3.4251773357391357, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.023399339988827705, + "learning_rate": 0.0003436596561623084, + "loss": 3.371610641479492, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.02192695252597332, + "learning_rate": 0.0003433955632325288, + "loss": 3.387253761291504, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.02241935208439827, + "learning_rate": 0.00034313143775677353, + "loss": 3.434523105621338, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.023202672600746155, + "learning_rate": 0.00034286727994784367, + "loss": 3.369771957397461, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.021333014592528343, + "learning_rate": 0.0003426030900185665, + "loss": 3.439004421234131, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.023988714441657066, + "learning_rate": 0.0003423388681817949, + "loss": 3.396838665008545, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.023885777220129967, + "learning_rate": 0.00034207461465040793, + "loss": 3.4114322662353516, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.02263713628053665, + "learning_rate": 0.0003418103296373096, + "loss": 3.4139351844787598, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.022404303774237633, + "learning_rate": 0.00034154601335542964, + "loss": 3.4200823307037354, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.021705076098442078, + "learning_rate": 0.00034128166601772304, + "loss": 3.4092609882354736, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.02273249812424183, + "learning_rate": 0.0003410172878371695, + "loss": 3.3931846618652344, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.022552402690052986, + "learning_rate": 0.00034075287902677394, + "loss": 3.396566867828369, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.023219924420118332, + "learning_rate": 0.0003404884397995655, + "loss": 3.369025707244873, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.022560900077223778, + "learning_rate": 0.00034022397036859837, + "loss": 3.4248225688934326, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.023370858281850815, + "learning_rate": 0.0003399594709469506, + "loss": 3.431628942489624, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.02469109185039997, + "learning_rate": 0.00033969494174772465, + "loss": 3.406843662261963, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.02417670376598835, + "learning_rate": 0.0003394303829840469, + "loss": 3.445046901702881, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.0237257182598114, + "learning_rate": 0.00033916579486906764, + "loss": 3.4490597248077393, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.02295520156621933, + "learning_rate": 0.0003389011776159607, + "loss": 3.4009876251220703, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.023158475756645203, + "learning_rate": 0.0003386365314379233, + "loss": 3.4383902549743652, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.023618478327989578, + "learning_rate": 0.0003383718565481763, + "loss": 3.421168804168701, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.022055329754948616, + "learning_rate": 0.0003381071531599633, + "loss": 3.382946729660034, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.021441346034407616, + "learning_rate": 0.00033784242148655115, + "loss": 3.4464712142944336, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.023031961172819138, + "learning_rate": 0.00033757766174122934, + "loss": 3.42844295501709, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.022832952439785004, + "learning_rate": 0.00033731287413731005, + "loss": 3.398231029510498, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.022094765678048134, + "learning_rate": 0.0003370480588881278, + "loss": 3.4270074367523193, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.0218376275151968, + "learning_rate": 0.0003367832162070395, + "loss": 3.4681742191314697, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.02364739589393139, + "learning_rate": 0.000336518346307424, + "loss": 3.431920051574707, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.023402225226163864, + "learning_rate": 0.0003362534494026824, + "loss": 3.4189023971557617, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.022775836288928986, + "learning_rate": 0.0003359885257062372, + "loss": 3.409396171569824, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.02445383556187153, + "learning_rate": 0.0003357235754315328, + "loss": 3.41017746925354, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.066352367401123, + "eval_runtime": 58.1658, + "eval_samples_per_second": 41.983, + "eval_steps_per_second": 1.324, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.02710847556591034, + "learning_rate": 0.0003354585987920345, + "loss": 3.3155131340026855, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.03428799286484718, + "learning_rate": 0.0003351935960012296, + "loss": 3.327408790588379, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.03434379771351814, + "learning_rate": 0.0003349285672726259, + "loss": 3.3282251358032227, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.03405233472585678, + "learning_rate": 0.0003346635128197522, + "loss": 3.2932448387145996, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.031194346025586128, + "learning_rate": 0.0003343984328561581, + "loss": 3.336615800857544, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.031816478818655014, + "learning_rate": 0.00033413332759541376, + "loss": 3.3375275135040283, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.03413299471139908, + "learning_rate": 0.0003338681972511098, + "loss": 3.3710694313049316, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.034200165420770645, + "learning_rate": 0.0003336030420368568, + "loss": 3.3248214721679688, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.030694160610437393, + "learning_rate": 0.0003333378621662857, + "loss": 3.3297901153564453, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.031172804534435272, + "learning_rate": 0.00033307265785304684, + "loss": 3.287217140197754, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.031914614140987396, + "learning_rate": 0.00033280742931081076, + "loss": 3.3218531608581543, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.0332055389881134, + "learning_rate": 0.00033254217675326737, + "loss": 3.3454785346984863, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.034956980496644974, + "learning_rate": 0.0003322769003941257, + "loss": 3.3193702697753906, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.03618055582046509, + "learning_rate": 0.00033201160044711423, + "loss": 3.3661017417907715, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.030322827398777008, + "learning_rate": 0.0003317462771259802, + "loss": 3.353989601135254, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.029768960550427437, + "learning_rate": 0.00033148093064449006, + "loss": 3.3254497051239014, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.028036782518029213, + "learning_rate": 0.0003312155612164284, + "loss": 3.3276920318603516, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.028774483129382133, + "learning_rate": 0.00033095016905559883, + "loss": 3.3458826541900635, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.030403515323996544, + "learning_rate": 0.0003306847543758227, + "loss": 3.3332793712615967, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.028468478471040726, + "learning_rate": 0.0003304193173909401, + "loss": 3.3167755603790283, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.029126476496458054, + "learning_rate": 0.00033015385831480873, + "loss": 3.299426794052124, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.030746638774871826, + "learning_rate": 0.0003298883773613043, + "loss": 3.3581809997558594, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.030241351574659348, + "learning_rate": 0.0003296228747443197, + "loss": 3.3415942192077637, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.029754415154457092, + "learning_rate": 0.0003293573506777659, + "loss": 3.3525662422180176, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.026896746829152107, + "learning_rate": 0.00032909180537557076, + "loss": 3.321808099746704, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.026361066848039627, + "learning_rate": 0.00032882623905167917, + "loss": 3.329258441925049, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.027747981250286102, + "learning_rate": 0.00032856065192005335, + "loss": 3.3774523735046387, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.026270100846886635, + "learning_rate": 0.00032829504419467194, + "loss": 3.3419976234436035, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.02731695957481861, + "learning_rate": 0.0003280294160895303, + "loss": 3.3068814277648926, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.026583299040794373, + "learning_rate": 0.0003277637678186402, + "loss": 3.303804874420166, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.026615949347615242, + "learning_rate": 0.00032749809959602973, + "loss": 3.3812599182128906, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.027221621945500374, + "learning_rate": 0.0003272324116357428, + "loss": 3.327211618423462, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.025025667622685432, + "learning_rate": 0.0003269667041518395, + "loss": 3.339163303375244, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.02593836560845375, + "learning_rate": 0.0003267009773583956, + "loss": 3.3517372608184814, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.023697076365351677, + "learning_rate": 0.00032643523146950234, + "loss": 3.379819869995117, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.02425902709364891, + "learning_rate": 0.0003261694666992664, + "loss": 3.390653610229492, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.025133250281214714, + "learning_rate": 0.0003259036832618096, + "loss": 3.364495277404785, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.025542905554175377, + "learning_rate": 0.0003256378813712688, + "loss": 3.3584814071655273, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.026041125878691673, + "learning_rate": 0.000325372061241796, + "loss": 3.36777400970459, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.02391037344932556, + "learning_rate": 0.00032510622308755746, + "loss": 3.3809385299682617, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.02379394695162773, + "learning_rate": 0.00032484036712273424, + "loss": 3.354257583618164, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.0234722550958395, + "learning_rate": 0.0003245744935615219, + "loss": 3.3736205101013184, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.024123650044202805, + "learning_rate": 0.0003243086026181296, + "loss": 3.3589959144592285, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.024013696238398552, + "learning_rate": 0.00032404269450678116, + "loss": 3.353874444961548, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.022617843002080917, + "learning_rate": 0.00032377676944171375, + "loss": 3.364056348800659, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.022730547934770584, + "learning_rate": 0.00032351082763717857, + "loss": 3.378067970275879, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.022257911041378975, + "learning_rate": 0.0003232448693074399, + "loss": 3.379342555999756, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.023247094824910164, + "learning_rate": 0.00032297889466677575, + "loss": 3.3537116050720215, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.022764021530747414, + "learning_rate": 0.000322712903929477, + "loss": 3.3969063758850098, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.02260403335094452, + "learning_rate": 0.0003224468973098477, + "loss": 3.3727574348449707, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.021772351115942, + "learning_rate": 0.0003221808750222044, + "loss": 3.335434675216675, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.02238459885120392, + "learning_rate": 0.0003219148372808766, + "loss": 3.3440277576446533, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.0230540968477726, + "learning_rate": 0.00032164878430020606, + "loss": 3.3997583389282227, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.0220821350812912, + "learning_rate": 0.00032138271629454684, + "loss": 3.377815008163452, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.021497417241334915, + "learning_rate": 0.00032111663347826505, + "loss": 3.3696370124816895, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.02163885347545147, + "learning_rate": 0.00032085053606573896, + "loss": 3.3489794731140137, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.022394269704818726, + "learning_rate": 0.0003205844242713584, + "loss": 3.3792812824249268, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.022390708327293396, + "learning_rate": 0.0003203182983095248, + "loss": 3.3733272552490234, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.023712158203125, + "learning_rate": 0.0003200521583946511, + "loss": 3.3716251850128174, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.02398860827088356, + "learning_rate": 0.0003197860047411613, + "loss": 3.381779670715332, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.02309161052107811, + "learning_rate": 0.00031951983756349066, + "loss": 3.3923277854919434, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.021400388330221176, + "learning_rate": 0.00031925365707608536, + "loss": 3.407397747039795, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.022534485906362534, + "learning_rate": 0.00031898746349340204, + "loss": 3.3740391731262207, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.023533202707767487, + "learning_rate": 0.0003187212570299082, + "loss": 3.3877944946289062, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.023398835211992264, + "learning_rate": 0.0003184550379000815, + "loss": 3.3936471939086914, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.020927201956510544, + "learning_rate": 0.00031818880631841005, + "loss": 3.339545249938965, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.021453995257616043, + "learning_rate": 0.0003179225624993915, + "loss": 3.3647522926330566, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.022847114130854607, + "learning_rate": 0.0003176563066575341, + "loss": 3.405144691467285, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.022766996175050735, + "learning_rate": 0.0003173900390073549, + "loss": 3.392810344696045, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.022103948518633842, + "learning_rate": 0.0003171237597633813, + "loss": 3.3973004817962646, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.022282181307673454, + "learning_rate": 0.0003168574691401496, + "loss": 3.409998893737793, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.025435522198677063, + "learning_rate": 0.0003165911673522053, + "loss": 3.423318386077881, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.078423500061035, + "eval_runtime": 57.3189, + "eval_samples_per_second": 42.604, + "eval_steps_per_second": 1.343, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.02624760940670967, + "learning_rate": 0.00031632485461410294, + "loss": 3.261288642883301, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.03552987053990364, + "learning_rate": 0.00031605853114040583, + "loss": 3.2682106494903564, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.040380869060754776, + "learning_rate": 0.00031579219714568613, + "loss": 3.2813782691955566, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.03861293941736221, + "learning_rate": 0.0003155258528445242, + "loss": 3.258193016052246, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.04091731086373329, + "learning_rate": 0.0003152594984515089, + "loss": 3.3052220344543457, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.043269723653793335, + "learning_rate": 0.00031499313418123704, + "loss": 3.2820217609405518, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.04580320790410042, + "learning_rate": 0.0003147267602483138, + "loss": 3.312005043029785, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.047336459159851074, + "learning_rate": 0.00031446037686735153, + "loss": 3.3217713832855225, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.04587360844016075, + "learning_rate": 0.0003141939842529706, + "loss": 3.2869949340820312, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.04002532362937927, + "learning_rate": 0.0003139275826197987, + "loss": 3.2879223823547363, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.0368693508207798, + "learning_rate": 0.00031366117218247094, + "loss": 3.2983577251434326, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.037317775189876556, + "learning_rate": 0.00031339475315562916, + "loss": 3.3062524795532227, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.03655700013041496, + "learning_rate": 0.00031312832575392267, + "loss": 3.31072735786438, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.033466216176748276, + "learning_rate": 0.0003128618901920071, + "loss": 3.2954514026641846, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.030785970389842987, + "learning_rate": 0.0003125954466845446, + "loss": 3.27952241897583, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.029669342562556267, + "learning_rate": 0.00031232899544620404, + "loss": 3.2613229751586914, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.033505845814943314, + "learning_rate": 0.00031206253669166045, + "loss": 3.2987258434295654, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.032395437359809875, + "learning_rate": 0.0003117960706355947, + "loss": 3.296708583831787, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.031322430819272995, + "learning_rate": 0.0003115295974926936, + "loss": 3.3160252571105957, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.029345186427235603, + "learning_rate": 0.00031126311747765005, + "loss": 3.307143211364746, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.026715675368905067, + "learning_rate": 0.00031099663080516186, + "loss": 3.286012649536133, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.028592750430107117, + "learning_rate": 0.0003107301376899327, + "loss": 3.2528645992279053, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.028861764818429947, + "learning_rate": 0.00031046363834667123, + "loss": 3.2795748710632324, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.02609528973698616, + "learning_rate": 0.00031019713299009123, + "loss": 3.326317071914673, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.027718454599380493, + "learning_rate": 0.000309930621834911, + "loss": 3.268110513687134, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.026382867246866226, + "learning_rate": 0.0003096641050958541, + "loss": 3.294499397277832, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.027083583176136017, + "learning_rate": 0.00030939758298764795, + "loss": 3.3202130794525146, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.02516467310488224, + "learning_rate": 0.00030913105572502483, + "loss": 3.3485021591186523, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.024148069322109222, + "learning_rate": 0.00030886452352272095, + "loss": 3.3099541664123535, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.024552857503294945, + "learning_rate": 0.00030859798659547624, + "loss": 3.2885942459106445, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.02551114559173584, + "learning_rate": 0.0003083314451580349, + "loss": 3.3232040405273438, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.025423647835850716, + "learning_rate": 0.0003080648994251445, + "loss": 3.3097832202911377, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.023211417719721794, + "learning_rate": 0.0003077983496115561, + "loss": 3.3075966835021973, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.024623898789286613, + "learning_rate": 0.00030753179593202406, + "loss": 3.2917377948760986, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.02677663415670395, + "learning_rate": 0.0003072652386013059, + "loss": 3.356265068054199, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.02673642337322235, + "learning_rate": 0.00030699867783416207, + "loss": 3.314903736114502, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.024466225877404213, + "learning_rate": 0.0003067321138453557, + "loss": 3.30912446975708, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.025370748713612556, + "learning_rate": 0.0003064655468496527, + "loss": 3.363640069961548, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.026536114513874054, + "learning_rate": 0.0003061989770618211, + "loss": 3.304851531982422, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.026536893099546432, + "learning_rate": 0.00030593240469663154, + "loss": 3.342902898788452, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.02645307406783104, + "learning_rate": 0.0003056658299688563, + "loss": 3.3164432048797607, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.02523060515522957, + "learning_rate": 0.00030539925309327017, + "loss": 3.34376859664917, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.02590431459248066, + "learning_rate": 0.0003051326742846491, + "loss": 3.3479061126708984, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.0230655986815691, + "learning_rate": 0.000304866093757771, + "loss": 3.2890982627868652, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.02403893694281578, + "learning_rate": 0.00030459951172741474, + "loss": 3.345456600189209, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.02238907851278782, + "learning_rate": 0.0003043329284083609, + "loss": 3.3471767902374268, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.021377569064497948, + "learning_rate": 0.00030406634401539053, + "loss": 3.319232225418091, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.02272353693842888, + "learning_rate": 0.00030379975876328614, + "loss": 3.316892623901367, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.02413824200630188, + "learning_rate": 0.00030353317286683053, + "loss": 3.3450968265533447, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.02334682270884514, + "learning_rate": 0.0003032665865408073, + "loss": 3.3640642166137695, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.021966520696878433, + "learning_rate": 0.000303, + "loss": 3.358555316925049, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.023984894156455994, + "learning_rate": 0.0003027334134591926, + "loss": 3.358398914337158, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.023520268499851227, + "learning_rate": 0.00030246682713316945, + "loss": 3.3224587440490723, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.0231285709887743, + "learning_rate": 0.0003022002412367138, + "loss": 3.3671562671661377, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.022818170487880707, + "learning_rate": 0.00030193365598460946, + "loss": 3.361027717590332, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.02280694991350174, + "learning_rate": 0.0003016670715916391, + "loss": 3.2923383712768555, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.021362345665693283, + "learning_rate": 0.00030140048827258524, + "loss": 3.3346734046936035, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.02372218668460846, + "learning_rate": 0.0003011339062422289, + "loss": 3.354748249053955, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.022949764505028725, + "learning_rate": 0.00030086732571535076, + "loss": 3.3857789039611816, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.021954255178570747, + "learning_rate": 0.0003006007469067297, + "loss": 3.343330144882202, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.022438595071434975, + "learning_rate": 0.0003003341700311436, + "loss": 3.328618049621582, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.021969925612211227, + "learning_rate": 0.0003000675953033685, + "loss": 3.355971574783325, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.022706981748342514, + "learning_rate": 0.0002998010229381789, + "loss": 3.3529577255249023, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.024083172902464867, + "learning_rate": 0.0002995344531503474, + "loss": 3.348170757293701, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.023259354755282402, + "learning_rate": 0.0002992678861546442, + "loss": 3.3474369049072266, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.02218157798051834, + "learning_rate": 0.00029900132216583786, + "loss": 3.369053840637207, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.02305331453680992, + "learning_rate": 0.0002987347613986941, + "loss": 3.3622641563415527, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.022796178236603737, + "learning_rate": 0.0002984682040679759, + "loss": 3.351510524749756, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.021299205720424652, + "learning_rate": 0.00029820165038844395, + "loss": 3.339406728744507, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.022786924615502357, + "learning_rate": 0.00029793510057485556, + "loss": 3.3182196617126465, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.021729128435254097, + "learning_rate": 0.00029766855484196513, + "loss": 3.383904218673706, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.024604283273220062, + "learning_rate": 0.0002974020134045238, + "loss": 3.2912864685058594, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.0894880294799805, + "eval_runtime": 58.3967, + "eval_samples_per_second": 41.817, + "eval_steps_per_second": 1.319, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.02673362009227276, + "learning_rate": 0.00029713547647727904, + "loss": 3.270986795425415, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.04036599025130272, + "learning_rate": 0.00029686894427497515, + "loss": 3.2358999252319336, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.047250088304281235, + "learning_rate": 0.0002966024170123519, + "loss": 3.2706189155578613, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.04642819985747337, + "learning_rate": 0.0002963358949041459, + "loss": 3.236340284347534, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.039405327290296555, + "learning_rate": 0.00029606937816508897, + "loss": 3.203136682510376, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.038748081773519516, + "learning_rate": 0.00029580286700990887, + "loss": 3.2636070251464844, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.039233580231666565, + "learning_rate": 0.00029553636165332876, + "loss": 3.2751896381378174, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.03429320082068443, + "learning_rate": 0.0002952698623100673, + "loss": 3.2765679359436035, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.03267038241028786, + "learning_rate": 0.0002950033691948382, + "loss": 3.2387681007385254, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.03349406272172928, + "learning_rate": 0.00029473688252234994, + "loss": 3.282162666320801, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.032002877444028854, + "learning_rate": 0.0002944704025073064, + "loss": 3.246623992919922, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.031509071588516235, + "learning_rate": 0.00029420392936440525, + "loss": 3.2374932765960693, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.02958357334136963, + "learning_rate": 0.00029393746330833953, + "loss": 3.24362850189209, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.03248635679483414, + "learning_rate": 0.0002936710045537959, + "loss": 3.237210273742676, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.03230655938386917, + "learning_rate": 0.0002934045533154554, + "loss": 3.251129150390625, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.029688173905014992, + "learning_rate": 0.00029313810980799296, + "loss": 3.2141575813293457, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.027773940935730934, + "learning_rate": 0.00029287167424607726, + "loss": 3.264249086380005, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.02687031403183937, + "learning_rate": 0.00029260524684437077, + "loss": 3.2513632774353027, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.02503623627126217, + "learning_rate": 0.00029233882781752905, + "loss": 3.217961072921753, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.027071228250861168, + "learning_rate": 0.00029207241738020136, + "loss": 3.2784645557403564, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.02600773423910141, + "learning_rate": 0.0002918060157470294, + "loss": 3.2641382217407227, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.026998162269592285, + "learning_rate": 0.00029153962313264856, + "loss": 3.2646782398223877, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.025579402223229408, + "learning_rate": 0.0002912732397516862, + "loss": 3.2700581550598145, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.02645333670079708, + "learning_rate": 0.00029100686581876283, + "loss": 3.2545266151428223, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.02717350795865059, + "learning_rate": 0.0002907405015484911, + "loss": 3.2795705795288086, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.02675914391875267, + "learning_rate": 0.0002904741471554758, + "loss": 3.2936904430389404, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.02751768007874489, + "learning_rate": 0.0002902078028543139, + "loss": 3.2329440116882324, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.025446882471442223, + "learning_rate": 0.00028994146885959416, + "loss": 3.274294376373291, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.02365768514573574, + "learning_rate": 0.00028967514538589715, + "loss": 3.2585229873657227, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.0266884732991457, + "learning_rate": 0.0002894088326477947, + "loss": 3.2872962951660156, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.02572469227015972, + "learning_rate": 0.0002891425308598503, + "loss": 3.253828287124634, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.02540273219347, + "learning_rate": 0.00028887624023661866, + "loss": 3.303926944732666, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.024694625288248062, + "learning_rate": 0.00028860996099264505, + "loss": 3.264145851135254, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.027233170345425606, + "learning_rate": 0.000288343693342466, + "loss": 3.286924362182617, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02635285072028637, + "learning_rate": 0.00028807743750060847, + "loss": 3.2930331230163574, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.025157805532217026, + "learning_rate": 0.00028781119368159004, + "loss": 3.2906906604766846, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.024729596450924873, + "learning_rate": 0.00028754496209991843, + "loss": 3.277128219604492, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.023809000849723816, + "learning_rate": 0.0002872787429700917, + "loss": 3.2957217693328857, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02474421076476574, + "learning_rate": 0.0002870125365065979, + "loss": 3.27791690826416, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.02439294196665287, + "learning_rate": 0.00028674634292391463, + "loss": 3.2912981510162354, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.023601334542036057, + "learning_rate": 0.0002864801624365093, + "loss": 3.317446231842041, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.0253264419734478, + "learning_rate": 0.0002862139952588387, + "loss": 3.2742302417755127, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.02389134280383587, + "learning_rate": 0.000285947841605349, + "loss": 3.281032085418701, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.02451813407242298, + "learning_rate": 0.00028568170169047523, + "loss": 3.3339295387268066, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.024324845522642136, + "learning_rate": 0.00028541557572864154, + "loss": 3.290402889251709, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.024939008057117462, + "learning_rate": 0.000285149463934261, + "loss": 3.2933621406555176, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.024450505152344704, + "learning_rate": 0.0002848833665217349, + "loss": 3.3055014610290527, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.02262270078063011, + "learning_rate": 0.0002846172837054532, + "loss": 3.2964205741882324, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.02311946265399456, + "learning_rate": 0.0002843512156997939, + "loss": 3.3102402687072754, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.024362705647945404, + "learning_rate": 0.0002840851627191234, + "loss": 3.3020925521850586, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.024785032495856285, + "learning_rate": 0.0002838191249777955, + "loss": 3.287290096282959, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.02358846180140972, + "learning_rate": 0.00028355310269015225, + "loss": 3.334261894226074, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.024330584332346916, + "learning_rate": 0.00028328709607052297, + "loss": 3.2799081802368164, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.023546984419226646, + "learning_rate": 0.0002830211053332242, + "loss": 3.3228847980499268, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.023060832172632217, + "learning_rate": 0.0002827551306925601, + "loss": 3.325793981552124, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.02388455718755722, + "learning_rate": 0.00028248917236282147, + "loss": 3.2892062664031982, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.023740701377391815, + "learning_rate": 0.0002822232305582863, + "loss": 3.2969369888305664, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.02458217367529869, + "learning_rate": 0.0002819573054932188, + "loss": 3.315610408782959, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.023173579946160316, + "learning_rate": 0.0002816913973818703, + "loss": 3.2561936378479004, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.024137936532497406, + "learning_rate": 0.0002814255064384781, + "loss": 3.2839202880859375, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.022436272352933884, + "learning_rate": 0.00028115963287726563, + "loss": 3.3057351112365723, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.023997878655791283, + "learning_rate": 0.0002808937769124425, + "loss": 3.3583154678344727, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.024119896814227104, + "learning_rate": 0.000280627938758204, + "loss": 3.322892427444458, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.022283365949988365, + "learning_rate": 0.00028036211862873117, + "loss": 3.3145735263824463, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.023433363065123558, + "learning_rate": 0.0002800963167381904, + "loss": 3.3564529418945312, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.02224746160209179, + "learning_rate": 0.00027983053330073355, + "loss": 3.2985916137695312, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.023786570876836777, + "learning_rate": 0.00027956476853049765, + "loss": 3.302539587020874, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.025200961157679558, + "learning_rate": 0.00027929902264160435, + "loss": 3.309976577758789, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.023849809542298317, + "learning_rate": 0.00027903329584816046, + "loss": 3.343660354614258, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.023311031982302666, + "learning_rate": 0.0002787675883642572, + "loss": 3.2978909015655518, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.023174891248345375, + "learning_rate": 0.00027850190040397036, + "loss": 3.326188802719116, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.02681034244596958, + "learning_rate": 0.00027823623218135976, + "loss": 3.3422017097473145, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.109009265899658, + "eval_runtime": 58.0506, + "eval_samples_per_second": 42.067, + "eval_steps_per_second": 1.326, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.028057295829057693, + "learning_rate": 0.0002779705839104697, + "loss": 3.1825037002563477, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.034786805510520935, + "learning_rate": 0.0002777049558053281, + "loss": 3.200389862060547, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.03751441463828087, + "learning_rate": 0.0002774393480799466, + "loss": 3.2047512531280518, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.035590436309576035, + "learning_rate": 0.00027717376094832076, + "loss": 3.1735124588012695, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.03426426649093628, + "learning_rate": 0.0002769081946244293, + "loss": 3.199483871459961, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.037030890583992004, + "learning_rate": 0.00027664264932223407, + "loss": 3.213322639465332, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.032468877732753754, + "learning_rate": 0.00027637712525568024, + "loss": 3.179354190826416, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.027202758938074112, + "learning_rate": 0.0002761116226386958, + "loss": 3.216036319732666, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.031184563413262367, + "learning_rate": 0.00027584614168519125, + "loss": 3.208205223083496, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.029414812102913857, + "learning_rate": 0.00027558068260905976, + "loss": 3.180370569229126, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.02679901383817196, + "learning_rate": 0.00027531524562417734, + "loss": 3.210325002670288, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.02863631211221218, + "learning_rate": 0.00027504983094440126, + "loss": 3.1601076126098633, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.029135366901755333, + "learning_rate": 0.0002747844387835716, + "loss": 3.2262887954711914, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.030535712838172913, + "learning_rate": 0.0002745190693555099, + "loss": 3.23160719871521, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.030563877895474434, + "learning_rate": 0.0002742537228740197, + "loss": 3.255211353302002, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.02906505949795246, + "learning_rate": 0.00027398839955288575, + "loss": 3.2006797790527344, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.030871668830513954, + "learning_rate": 0.00027372309960587424, + "loss": 3.2442867755889893, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.02913421392440796, + "learning_rate": 0.0002734578232467327, + "loss": 3.218442440032959, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.030087953433394432, + "learning_rate": 0.0002731925706891891, + "loss": 3.2164535522460938, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.034147944301366806, + "learning_rate": 0.0002729273421469532, + "loss": 3.218660831451416, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.03593966364860535, + "learning_rate": 0.00027266213783371434, + "loss": 3.2240614891052246, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.032090652734041214, + "learning_rate": 0.00027239695796314327, + "loss": 3.2309441566467285, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.02776424027979374, + "learning_rate": 0.0002721318027488902, + "loss": 3.2702016830444336, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.027174271643161774, + "learning_rate": 0.00027186667240458617, + "loss": 3.19810152053833, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.028479408472776413, + "learning_rate": 0.0002716015671438419, + "loss": 3.2178421020507812, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.026820644736289978, + "learning_rate": 0.0002713364871802478, + "loss": 3.2634482383728027, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.026427917182445526, + "learning_rate": 0.0002710714327273741, + "loss": 3.2452914714813232, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.02679610252380371, + "learning_rate": 0.0002708064039987704, + "loss": 3.243825912475586, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.026115626096725464, + "learning_rate": 0.00027054140120796546, + "loss": 3.218787670135498, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.026686571538448334, + "learning_rate": 0.0002702764245684672, + "loss": 3.2785303592681885, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.02685859426856041, + "learning_rate": 0.0002700114742937627, + "loss": 3.2572226524353027, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.026968171820044518, + "learning_rate": 0.00026974655059731754, + "loss": 3.2625553607940674, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.025397246703505516, + "learning_rate": 0.0002694816536925759, + "loss": 3.262303352355957, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.025660471990704536, + "learning_rate": 0.00026921678379296057, + "loss": 3.269401788711548, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.02713891491293907, + "learning_rate": 0.0002689519411118722, + "loss": 3.2416529655456543, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.027591092512011528, + "learning_rate": 0.00026868712586269, + "loss": 3.2814316749572754, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.024462848901748657, + "learning_rate": 0.00026842233825877064, + "loss": 3.249725103378296, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.024868478998541832, + "learning_rate": 0.0002681575785134487, + "loss": 3.231574535369873, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.027015401050448418, + "learning_rate": 0.0002678928468400366, + "loss": 3.282627582550049, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.02678140625357628, + "learning_rate": 0.00026762814345182367, + "loss": 3.2208900451660156, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.026830527931451797, + "learning_rate": 0.0002673634685620767, + "loss": 3.260575294494629, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.02430890128016472, + "learning_rate": 0.0002670988223840393, + "loss": 3.2522997856140137, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.027047613635659218, + "learning_rate": 0.0002668342051309324, + "loss": 3.2721078395843506, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.026861369609832764, + "learning_rate": 0.00026656961701595306, + "loss": 3.2235946655273438, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.026650549843907356, + "learning_rate": 0.0002663050582522753, + "loss": 3.302743434906006, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.02551819011569023, + "learning_rate": 0.00026604052905304946, + "loss": 3.265763282775879, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.024843450635671616, + "learning_rate": 0.00026577602963140156, + "loss": 3.272397518157959, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.024423206225037575, + "learning_rate": 0.0002655115602004345, + "loss": 3.253495454788208, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.024180661886930466, + "learning_rate": 0.00026524712097322604, + "loss": 3.260438919067383, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.023554982617497444, + "learning_rate": 0.0002649827121628305, + "loss": 3.2342851161956787, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.023470086976885796, + "learning_rate": 0.00026471833398227694, + "loss": 3.251622438430786, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.024405231699347496, + "learning_rate": 0.0002644539866445703, + "loss": 3.2588815689086914, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.024147814139723778, + "learning_rate": 0.00026418967036269045, + "loss": 3.2723565101623535, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.023217910900712013, + "learning_rate": 0.0002639253853495921, + "loss": 3.2555792331695557, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.024921158328652382, + "learning_rate": 0.00026366113181820506, + "loss": 3.26884126663208, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.02455919235944748, + "learning_rate": 0.00026339690998143346, + "loss": 3.2674927711486816, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.024725932627916336, + "learning_rate": 0.00026313272005215637, + "loss": 3.2511110305786133, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.025769414380192757, + "learning_rate": 0.00026286856224322645, + "loss": 3.290651321411133, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.025928806513547897, + "learning_rate": 0.00026260443676747114, + "loss": 3.302396297454834, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.02499072253704071, + "learning_rate": 0.00026234034383769157, + "loss": 3.2813124656677246, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.025431277230381966, + "learning_rate": 0.0002620762836666625, + "loss": 3.234684705734253, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.026871860027313232, + "learning_rate": 0.00026181225646713257, + "loss": 3.2800331115722656, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.02578776702284813, + "learning_rate": 0.0002615482624518236, + "loss": 3.3117146492004395, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.026318378746509552, + "learning_rate": 0.0002612843018334309, + "loss": 3.257615089416504, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.025542527437210083, + "learning_rate": 0.00026102037482462256, + "loss": 3.2969868183135986, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.025723429396748543, + "learning_rate": 0.0002607564816380403, + "loss": 3.290259838104248, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.024588774889707565, + "learning_rate": 0.0002604926224862975, + "loss": 3.284090280532837, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.024619122967123985, + "learning_rate": 0.0002602287975819811, + "loss": 3.284879207611084, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.024441618472337723, + "learning_rate": 0.00025996500713765, + "loss": 3.283104658126831, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.02402760274708271, + "learning_rate": 0.0002597012513658352, + "loss": 3.3013617992401123, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.025028521195054054, + "learning_rate": 0.00025943753047904026, + "loss": 3.2892568111419678, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.029445001855492592, + "learning_rate": 0.0002591738446897401, + "loss": 3.2474446296691895, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.128176689147949, + "eval_runtime": 58.8588, + "eval_samples_per_second": 41.489, + "eval_steps_per_second": 1.308, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.0298826452344656, + "learning_rate": 0.00025891019421038177, + "loss": 3.138103485107422, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.0387548953294754, + "learning_rate": 0.00025864657925338345, + "loss": 3.161492347717285, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.03816070780158043, + "learning_rate": 0.00025838300003113514, + "loss": 3.1739673614501953, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.03409738838672638, + "learning_rate": 0.0002581194567559979, + "loss": 3.182478189468384, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.033135563135147095, + "learning_rate": 0.00025785594964030365, + "loss": 3.1720199584960938, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.03203624114394188, + "learning_rate": 0.0002575924788963554, + "loss": 3.1699862480163574, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.031237754970788956, + "learning_rate": 0.0002573290447364268, + "loss": 3.179363250732422, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.03342146426439285, + "learning_rate": 0.0002570656473727619, + "loss": 3.1822280883789062, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.03561089560389519, + "learning_rate": 0.0002568022870175752, + "loss": 3.1767094135284424, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.034710463136434555, + "learning_rate": 0.00025653896388305134, + "loss": 3.218817949295044, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.030870331451296806, + "learning_rate": 0.0002562756781813453, + "loss": 3.168602705001831, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.028004391118884087, + "learning_rate": 0.00025601243012458126, + "loss": 3.224155902862549, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.030953701585531235, + "learning_rate": 0.00025574921992485366, + "loss": 3.1789557933807373, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.030289432033896446, + "learning_rate": 0.000255486047794226, + "loss": 3.1890387535095215, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.030684132128953934, + "learning_rate": 0.0002552229139447316, + "loss": 3.195427417755127, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.03022196516394615, + "learning_rate": 0.00025495981858837235, + "loss": 3.1984965801239014, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.026935536414384842, + "learning_rate": 0.0002546967619371195, + "loss": 3.2014272212982178, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.02702922560274601, + "learning_rate": 0.0002544337442029133, + "loss": 3.205538034439087, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.026682600378990173, + "learning_rate": 0.00025417076559766194, + "loss": 3.1954591274261475, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.0273384228348732, + "learning_rate": 0.000253907826333243, + "loss": 3.1782338619232178, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.028587577864527702, + "learning_rate": 0.00025364492662150145, + "loss": 3.199565887451172, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.028848132118582726, + "learning_rate": 0.0002533820666742514, + "loss": 3.181786298751831, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.027535341680049896, + "learning_rate": 0.00025311924670327364, + "loss": 3.2013628482818604, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.02756025455892086, + "learning_rate": 0.00025285646692031804, + "loss": 3.1749587059020996, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.027358995750546455, + "learning_rate": 0.0002525937275371013, + "loss": 3.161881923675537, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.025918981060385704, + "learning_rate": 0.00025233102876530785, + "loss": 3.225224256515503, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.026479976251721382, + "learning_rate": 0.0002520683708165894, + "loss": 3.2129316329956055, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.02685079537332058, + "learning_rate": 0.0002518057539025646, + "loss": 3.1651554107666016, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.027378706261515617, + "learning_rate": 0.00025154317823481917, + "loss": 3.193108558654785, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.027240974828600883, + "learning_rate": 0.0002512806440249056, + "loss": 3.190605640411377, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.028065402060747147, + "learning_rate": 0.00025101815148434296, + "loss": 3.2068259716033936, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.028329014778137207, + "learning_rate": 0.00025075570082461684, + "loss": 3.2316060066223145, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.028920065611600876, + "learning_rate": 0.0002504932922571789, + "loss": 3.1933932304382324, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.028299354016780853, + "learning_rate": 0.00025023092599344706, + "loss": 3.192711353302002, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.025996020063757896, + "learning_rate": 0.0002499686022448049, + "loss": 3.243171453475952, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.0264595914632082, + "learning_rate": 0.00024970632122260216, + "loss": 3.199758529663086, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.02704528719186783, + "learning_rate": 0.0002494440831381538, + "loss": 3.2390713691711426, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.026571933180093765, + "learning_rate": 0.00024918188820274016, + "loss": 3.219440460205078, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.025562983006238937, + "learning_rate": 0.0002489197366276073, + "loss": 3.2538228034973145, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.024683184921741486, + "learning_rate": 0.0002486576286239657, + "loss": 3.2372591495513916, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.025656528770923615, + "learning_rate": 0.0002483955644029913, + "loss": 3.250558853149414, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.024807540699839592, + "learning_rate": 0.0002481335441758243, + "loss": 3.1856563091278076, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.026149194687604904, + "learning_rate": 0.0002478715681535699, + "loss": 3.172186851501465, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.02617705799639225, + "learning_rate": 0.00024760963654729717, + "loss": 3.250445604324341, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.026542512699961662, + "learning_rate": 0.0002473477495680397, + "loss": 3.220273494720459, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.024963509291410446, + "learning_rate": 0.00024708590742679546, + "loss": 3.2330946922302246, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.027144264429807663, + "learning_rate": 0.00024682411033452546, + "loss": 3.1860971450805664, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.02665683813393116, + "learning_rate": 0.0002465623585021553, + "loss": 3.257209062576294, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.024872658774256706, + "learning_rate": 0.0002463006521405733, + "loss": 3.2266182899475098, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.025235185399651527, + "learning_rate": 0.0002460389914606319, + "loss": 3.2134504318237305, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.025029495358467102, + "learning_rate": 0.00024577737667314615, + "loss": 3.2150087356567383, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.025630896911025047, + "learning_rate": 0.00024551580798889446, + "loss": 3.282879114151001, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.02722341939806938, + "learning_rate": 0.00024525428561861804, + "loss": 3.2366206645965576, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.02522372268140316, + "learning_rate": 0.0002449928097730207, + "loss": 3.2840523719787598, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.02494790591299534, + "learning_rate": 0.00024473138066276876, + "loss": 3.2405319213867188, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.026662737131118774, + "learning_rate": 0.0002444699984984909, + "loss": 3.239692211151123, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.026082608848810196, + "learning_rate": 0.0002442086634907782, + "loss": 3.2156195640563965, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.02423974685370922, + "learning_rate": 0.00024394737585018328, + "loss": 3.1863760948181152, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.025226688012480736, + "learning_rate": 0.00024368613578722113, + "loss": 3.20874285697937, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.02428225241601467, + "learning_rate": 0.00024342494351236799, + "loss": 3.2209601402282715, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.023819707334041595, + "learning_rate": 0.0002431637992360618, + "loss": 3.2564258575439453, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.025661710649728775, + "learning_rate": 0.00024290270316870184, + "loss": 3.244133949279785, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.02423049882054329, + "learning_rate": 0.0002426416555206484, + "loss": 3.2376630306243896, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.02382473088800907, + "learning_rate": 0.000242380656502223, + "loss": 3.2951290607452393, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.02488875202834606, + "learning_rate": 0.00024211970632370756, + "loss": 3.2848310470581055, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.02575223706662655, + "learning_rate": 0.0002418588051953453, + "loss": 3.2574503421783447, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.02462848648428917, + "learning_rate": 0.0002415979533273392, + "loss": 3.2464442253112793, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.024912388995289803, + "learning_rate": 0.0002413371509298531, + "loss": 3.2459192276000977, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.025020884349942207, + "learning_rate": 0.00024107639821301078, + "loss": 3.213622808456421, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.026330886408686638, + "learning_rate": 0.00024081569538689587, + "loss": 3.2305431365966797, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.02708602324128151, + "learning_rate": 0.0002405550426615521, + "loss": 3.2487688064575195, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.030908891931176186, + "learning_rate": 0.00024029444024698248, + "loss": 3.2110159397125244, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.146673202514648, + "eval_runtime": 58.7379, + "eval_samples_per_second": 41.574, + "eval_steps_per_second": 1.311, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.02919941022992134, + "learning_rate": 0.00024003388835314981, + "loss": 3.1533267498016357, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.03379782289266586, + "learning_rate": 0.00023977338718997595, + "loss": 3.1436538696289062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.033698830753564835, + "learning_rate": 0.00023951293696734195, + "loss": 3.15297532081604, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.03077736124396324, + "learning_rate": 0.00023925253789508806, + "loss": 3.128848075866699, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.031112324446439743, + "learning_rate": 0.00023899219018301286, + "loss": 3.1461119651794434, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.03346852958202362, + "learning_rate": 0.00023873189404087405, + "loss": 3.113943099975586, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.03307225927710533, + "learning_rate": 0.00023847164967838726, + "loss": 3.147498607635498, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.030230754986405373, + "learning_rate": 0.00023821145730522692, + "loss": 3.1162924766540527, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.02982981689274311, + "learning_rate": 0.0002379513171310251, + "loss": 3.144481658935547, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.031735632568597794, + "learning_rate": 0.00023769122936537207, + "loss": 3.133862018585205, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.0312766470015049, + "learning_rate": 0.00023743119421781614, + "loss": 3.110287666320801, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.029902562499046326, + "learning_rate": 0.00023717121189786256, + "loss": 3.1384904384613037, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.0306564774364233, + "learning_rate": 0.00023691128261497463, + "loss": 3.1138200759887695, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.03084711916744709, + "learning_rate": 0.0002366514065785725, + "loss": 3.155240535736084, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.03183220326900482, + "learning_rate": 0.00023639158399803382, + "loss": 3.1593515872955322, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.031826410442590714, + "learning_rate": 0.0002361318150826927, + "loss": 3.1517653465270996, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.030141707509756088, + "learning_rate": 0.0002358721000418404, + "loss": 3.1291589736938477, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.03158571943640709, + "learning_rate": 0.00023561243908472472, + "loss": 3.178101062774658, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.03510458394885063, + "learning_rate": 0.00023535283242054954, + "loss": 3.138767719268799, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.03660859912633896, + "learning_rate": 0.00023509328025847565, + "loss": 3.1708762645721436, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.034004807472229004, + "learning_rate": 0.00023483378280761907, + "loss": 3.1404073238372803, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.03077496401965618, + "learning_rate": 0.0002345743402770527, + "loss": 3.159081220626831, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.030384650453925133, + "learning_rate": 0.0002343149528758042, + "loss": 3.1872847080230713, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.02957277186214924, + "learning_rate": 0.00023405562081285766, + "loss": 3.165973424911499, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.029529055580496788, + "learning_rate": 0.0002337963442971522, + "loss": 3.123840808868408, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.027776138857007027, + "learning_rate": 0.00023353712353758212, + "loss": 3.1705269813537598, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.028943249955773354, + "learning_rate": 0.00023327795874299704, + "loss": 3.1181654930114746, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.028178401291370392, + "learning_rate": 0.0002330188501222012, + "loss": 3.201678991317749, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.02865665778517723, + "learning_rate": 0.0002327597978839538, + "loss": 3.18479061126709, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.029173649847507477, + "learning_rate": 0.00023250080223696852, + "loss": 3.1844916343688965, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.0280170775949955, + "learning_rate": 0.00023224186338991352, + "loss": 3.187105417251587, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.029103878885507584, + "learning_rate": 0.0002319829815514112, + "loss": 3.183281660079956, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.029045486822724342, + "learning_rate": 0.00023172415693003789, + "loss": 3.1848196983337402, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.028364380821585655, + "learning_rate": 0.000231465389734324, + "loss": 3.197348117828369, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.029651399701833725, + "learning_rate": 0.00023120668017275346, + "loss": 3.1429920196533203, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.0281438659876585, + "learning_rate": 0.00023094802845376403, + "loss": 3.1916513442993164, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.029437677934765816, + "learning_rate": 0.00023068943478574662, + "loss": 3.172391176223755, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.03433044254779816, + "learning_rate": 0.00023043089937704541, + "loss": 3.2296864986419678, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.03259492665529251, + "learning_rate": 0.00023017242243595796, + "loss": 3.1962363719940186, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.0299630519002676, + "learning_rate": 0.00022991400417073427, + "loss": 3.169673442840576, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.028495769947767258, + "learning_rate": 0.00022965564478957739, + "loss": 3.1855409145355225, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.0290420800447464, + "learning_rate": 0.0002293973445006427, + "loss": 3.23667573928833, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.03340664133429527, + "learning_rate": 0.00022913910351203816, + "loss": 3.1873788833618164, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.03253626823425293, + "learning_rate": 0.00022888092203182377, + "loss": 3.204796552658081, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.028142794966697693, + "learning_rate": 0.00022862280026801182, + "loss": 3.193922281265259, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.028805622830986977, + "learning_rate": 0.00022836473842856623, + "loss": 3.1863901615142822, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.030033612623810768, + "learning_rate": 0.00022810673672140272, + "loss": 3.198482036590576, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.028154220432043076, + "learning_rate": 0.0002278487953543889, + "loss": 3.215069055557251, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.028098132461309433, + "learning_rate": 0.00022759091453534308, + "loss": 3.2197628021240234, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.028679659590125084, + "learning_rate": 0.00022733309447203553, + "loss": 3.2062196731567383, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.029104258865118027, + "learning_rate": 0.00022707533537218695, + "loss": 3.2182908058166504, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.029629778116941452, + "learning_rate": 0.00022681763744346938, + "loss": 3.2165639400482178, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.028304366394877434, + "learning_rate": 0.00022656000089350524, + "loss": 3.1783785820007324, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.026900488883256912, + "learning_rate": 0.00022630242592986766, + "loss": 3.1876602172851562, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.027047287672758102, + "learning_rate": 0.0002260449127600803, + "loss": 3.1879465579986572, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.027424108237028122, + "learning_rate": 0.0002257874615916166, + "loss": 3.2407851219177246, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.029319508001208305, + "learning_rate": 0.00022553007263190067, + "loss": 3.1980130672454834, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.027114372700452805, + "learning_rate": 0.00022527274608830574, + "loss": 3.211357593536377, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.02579840086400509, + "learning_rate": 0.0002250154821681555, + "loss": 3.2067861557006836, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.02767331339418888, + "learning_rate": 0.0002247582810787227, + "loss": 3.198587417602539, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.026645265519618988, + "learning_rate": 0.00022450114302722957, + "loss": 3.177441120147705, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.026181906461715698, + "learning_rate": 0.0002242440682208477, + "loss": 3.226815700531006, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.027218692004680634, + "learning_rate": 0.00022398705686669748, + "loss": 3.1731510162353516, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.027551919221878052, + "learning_rate": 0.0002237301091718485, + "loss": 3.2268621921539307, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.025782009586691856, + "learning_rate": 0.00022347322534331865, + "loss": 3.2052621841430664, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.024231620132923126, + "learning_rate": 0.00022321640558807488, + "loss": 3.201770544052124, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.02474607340991497, + "learning_rate": 0.00022295965011303188, + "loss": 3.1935834884643555, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.026299327611923218, + "learning_rate": 0.00022270295912505317, + "loss": 3.223275899887085, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.02439914271235466, + "learning_rate": 0.00022244633283095, + "loss": 3.2513954639434814, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.02650653012096882, + "learning_rate": 0.0002221897714374814, + "loss": 3.1914851665496826, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.024847298860549927, + "learning_rate": 0.00022193327515135445, + "loss": 3.2173104286193848, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.028475770726799965, + "learning_rate": 0.00022167684417922337, + "loss": 3.2184863090515137, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.168707847595215, + "eval_runtime": 57.4035, + "eval_samples_per_second": 42.541, + "eval_steps_per_second": 1.341, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.02837727963924408, + "learning_rate": 0.00022142047872769014, + "loss": 3.119239330291748, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.03562312573194504, + "learning_rate": 0.00022116417900330365, + "loss": 3.092813014984131, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.0344194732606411, + "learning_rate": 0.00022090794521255997, + "loss": 3.0802855491638184, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.03188958019018173, + "learning_rate": 0.00022065177756190214, + "loss": 3.1177961826324463, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.032181013375520706, + "learning_rate": 0.00022039567625771968, + "loss": 3.1343281269073486, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.031085725873708725, + "learning_rate": 0.0002201396415063489, + "loss": 3.118828296661377, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.031995322555303574, + "learning_rate": 0.00021988367351407227, + "loss": 3.110109329223633, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.02949182130396366, + "learning_rate": 0.00021962777248711872, + "loss": 3.0972743034362793, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.029299668967723846, + "learning_rate": 0.00021937193863166292, + "loss": 3.1257238388061523, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.030558617785573006, + "learning_rate": 0.00021911617215382564, + "loss": 3.1216022968292236, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.03181575983762741, + "learning_rate": 0.00021886047325967348, + "loss": 3.093721866607666, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.030934303998947144, + "learning_rate": 0.00021860484215521805, + "loss": 3.0910396575927734, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.030235569924116135, + "learning_rate": 0.00021834927904641708, + "loss": 3.133371114730835, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.03350163996219635, + "learning_rate": 0.00021809378413917296, + "loss": 3.1178882122039795, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.03468719869852066, + "learning_rate": 0.00021783835763933344, + "loss": 3.1222469806671143, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.034203238785266876, + "learning_rate": 0.0002175829997526909, + "loss": 3.12453031539917, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.03321516141295433, + "learning_rate": 0.0002173277106849826, + "loss": 3.115428924560547, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.03298434242606163, + "learning_rate": 0.0002170724906418905, + "loss": 3.127288341522217, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.029987169429659843, + "learning_rate": 0.00021681733982904046, + "loss": 3.1194515228271484, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.031271349638700485, + "learning_rate": 0.00021656225845200332, + "loss": 3.1284165382385254, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.032201699912548065, + "learning_rate": 0.0002163072467162931, + "loss": 3.1912736892700195, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.0299740731716156, + "learning_rate": 0.00021605230482736845, + "loss": 3.1233785152435303, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.029613154008984566, + "learning_rate": 0.00021579743299063122, + "loss": 3.1535067558288574, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.03303765505552292, + "learning_rate": 0.00021554263141142726, + "loss": 3.1550910472869873, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.032644059509038925, + "learning_rate": 0.00021528790029504548, + "loss": 3.1367573738098145, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.031695179641246796, + "learning_rate": 0.00021503323984671814, + "loss": 3.1350858211517334, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.030685821548104286, + "learning_rate": 0.00021477865027162067, + "loss": 3.121593952178955, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.03091946430504322, + "learning_rate": 0.0002145241317748711, + "loss": 3.113060235977173, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.0282905176281929, + "learning_rate": 0.0002142696845615306, + "loss": 3.184551239013672, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.030020160600543022, + "learning_rate": 0.0002140153088366024, + "loss": 3.1673004627227783, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.031114062294363976, + "learning_rate": 0.00021376100480503276, + "loss": 3.162588596343994, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.03007666766643524, + "learning_rate": 0.00021350677267170973, + "loss": 3.132836103439331, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.029013793915510178, + "learning_rate": 0.00021325261264146342, + "loss": 3.1552720069885254, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.02784561738371849, + "learning_rate": 0.00021299852491906616, + "loss": 3.189272165298462, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.02759280428290367, + "learning_rate": 0.00021274450970923167, + "loss": 3.141814708709717, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.02699868381023407, + "learning_rate": 0.0002124905672166156, + "loss": 3.1127257347106934, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.02782190404832363, + "learning_rate": 0.0002122366976458146, + "loss": 3.193972587585449, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.027298318222165108, + "learning_rate": 0.00021198290120136702, + "loss": 3.1393446922302246, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.027321161702275276, + "learning_rate": 0.0002117291780877519, + "loss": 3.0926666259765625, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.026399722322821617, + "learning_rate": 0.00021147552850938941, + "loss": 3.1578259468078613, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.02770969271659851, + "learning_rate": 0.0002112219526706406, + "loss": 3.1816914081573486, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.028344258666038513, + "learning_rate": 0.00021096845077580675, + "loss": 3.1860158443450928, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.028091782703995705, + "learning_rate": 0.0002107150230291299, + "loss": 3.1585092544555664, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.026564039289951324, + "learning_rate": 0.00021046166963479204, + "loss": 3.131692409515381, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.026349127292633057, + "learning_rate": 0.00021020839079691553, + "loss": 3.160431385040283, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.030081138014793396, + "learning_rate": 0.00020995518671956256, + "loss": 3.1647543907165527, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.029250919818878174, + "learning_rate": 0.00020970205760673493, + "loss": 3.1522226333618164, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.0282477755099535, + "learning_rate": 0.00020944900366237453, + "loss": 3.139249324798584, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.0284534003585577, + "learning_rate": 0.00020919602509036195, + "loss": 3.154621124267578, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.02677394263446331, + "learning_rate": 0.0002089431220945178, + "loss": 3.1375081539154053, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.026511570438742638, + "learning_rate": 0.00020869029487860126, + "loss": 3.1355676651000977, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.02766270563006401, + "learning_rate": 0.00020843754364631068, + "loss": 3.143874168395996, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.027013977989554405, + "learning_rate": 0.0002081848686012832, + "loss": 3.1602487564086914, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.026629449799656868, + "learning_rate": 0.00020793226994709446, + "loss": 3.1549456119537354, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.0272684209048748, + "learning_rate": 0.0002076797478872588, + "loss": 3.1739020347595215, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.027645915746688843, + "learning_rate": 0.00020742730262522844, + "loss": 3.172959327697754, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.025413552299141884, + "learning_rate": 0.00020717493436439433, + "loss": 3.1649041175842285, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.028406929224729538, + "learning_rate": 0.00020692264330808455, + "loss": 3.1821887493133545, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.027753977105021477, + "learning_rate": 0.00020667042965956587, + "loss": 3.163238048553467, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.02693209983408451, + "learning_rate": 0.00020641829362204195, + "loss": 3.1921019554138184, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.02671878971159458, + "learning_rate": 0.00020616623539865446, + "loss": 3.159438133239746, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.027099428698420525, + "learning_rate": 0.00020591425519248213, + "loss": 3.1806793212890625, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.025098824873566628, + "learning_rate": 0.00020566235320654069, + "loss": 3.158310651779175, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.02655479498207569, + "learning_rate": 0.00020541052964378323, + "loss": 3.1927309036254883, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.025836164131760597, + "learning_rate": 0.00020515878470709928, + "loss": 3.1578540802001953, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.02657160349190235, + "learning_rate": 0.00020490711859931545, + "loss": 3.1658592224121094, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.027223270386457443, + "learning_rate": 0.00020465553152319418, + "loss": 3.1760241985321045, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.025695348158478737, + "learning_rate": 0.00020440402368143492, + "loss": 3.202389717102051, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.025819813832640648, + "learning_rate": 0.00020415259527667299, + "loss": 3.2111103534698486, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.02621106244623661, + "learning_rate": 0.0002039012465114796, + "loss": 3.1560299396514893, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.026585502550005913, + "learning_rate": 0.00020364997758836193, + "loss": 3.176602840423584, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.03038652054965496, + "learning_rate": 0.00020339878870976287, + "loss": 3.1799943447113037, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.193845748901367, + "eval_runtime": 57.9911, + "eval_samples_per_second": 42.11, + "eval_steps_per_second": 1.328, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.03135736659169197, + "learning_rate": 0.00020314768007806077, + "loss": 3.070986270904541, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.038986921310424805, + "learning_rate": 0.00020289665189556923, + "loss": 3.0708835124969482, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.035115309059619904, + "learning_rate": 0.00020264570436453707, + "loss": 3.1070001125335693, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.032868221402168274, + "learning_rate": 0.00020239483768714843, + "loss": 3.082258701324463, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.03703184798359871, + "learning_rate": 0.0002021440520655219, + "loss": 3.057744264602661, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.03831101953983307, + "learning_rate": 0.00020189334770171098, + "loss": 3.0789475440979004, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.036301176995038986, + "learning_rate": 0.00020164272479770352, + "loss": 3.0565457344055176, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.03166409209370613, + "learning_rate": 0.00020139218355542203, + "loss": 3.0574910640716553, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.02984458953142166, + "learning_rate": 0.000201141724176723, + "loss": 3.080018997192383, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.034001126885414124, + "learning_rate": 0.0002008913468633969, + "loss": 3.0813302993774414, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.03814122825860977, + "learning_rate": 0.00020064105181716847, + "loss": 3.0579185485839844, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.03870106488466263, + "learning_rate": 0.00020039083923969564, + "loss": 3.0857529640197754, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.04005642607808113, + "learning_rate": 0.0002001407093325704, + "loss": 3.0880727767944336, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.03760755434632301, + "learning_rate": 0.0001998906622973177, + "loss": 3.0795767307281494, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.03260907158255577, + "learning_rate": 0.0001996406983353961, + "loss": 3.1006722450256348, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.029866183176636696, + "learning_rate": 0.0001993908176481969, + "loss": 3.0850343704223633, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.030988436192274094, + "learning_rate": 0.0001991410204370445, + "loss": 3.0743892192840576, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.03698134422302246, + "learning_rate": 0.00019889130690319608, + "loss": 3.056002378463745, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.03696191683411598, + "learning_rate": 0.00019864167724784117, + "loss": 3.069124698638916, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.03453022986650467, + "learning_rate": 0.00019839213167210213, + "loss": 3.1017374992370605, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.032458193600177765, + "learning_rate": 0.00019814267037703294, + "loss": 3.1166419982910156, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.02892368845641613, + "learning_rate": 0.0001978932935636205, + "loss": 3.0863287448883057, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.030625877901911736, + "learning_rate": 0.00019764400143278276, + "loss": 3.0938801765441895, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.030280394479632378, + "learning_rate": 0.00019739479418537018, + "loss": 3.086432933807373, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.029181072488427162, + "learning_rate": 0.00019714567202216453, + "loss": 3.0977859497070312, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.028039127588272095, + "learning_rate": 0.00019689663514387896, + "loss": 3.112367868423462, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.02910272404551506, + "learning_rate": 0.00019664768375115798, + "loss": 3.1253533363342285, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.03086872771382332, + "learning_rate": 0.00019639881804457723, + "loss": 3.103689670562744, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.02883610874414444, + "learning_rate": 0.00019615003822464342, + "loss": 3.111599922180176, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.028490116819739342, + "learning_rate": 0.00019590134449179371, + "loss": 3.1410303115844727, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.028714407235383987, + "learning_rate": 0.00019565273704639648, + "loss": 3.133023738861084, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.029553238302469254, + "learning_rate": 0.00019540421608874994, + "loss": 3.107003688812256, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.02997349575161934, + "learning_rate": 0.0001951557818190831, + "loss": 3.135680675506592, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.029087591916322708, + "learning_rate": 0.00019490743443755504, + "loss": 3.1103882789611816, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.029947761446237564, + "learning_rate": 0.0001946591741442546, + "loss": 3.0961623191833496, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.02950555644929409, + "learning_rate": 0.00019441100113920076, + "loss": 3.1162495613098145, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.029569562524557114, + "learning_rate": 0.00019416291562234192, + "loss": 3.096372127532959, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.031366728246212006, + "learning_rate": 0.00019391491779355608, + "loss": 3.120089530944824, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.0313473716378212, + "learning_rate": 0.00019366700785265065, + "loss": 3.100785732269287, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.028021851554512978, + "learning_rate": 0.0001934191859993622, + "loss": 3.0940558910369873, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.027477117255330086, + "learning_rate": 0.00019317145243335642, + "loss": 3.1210484504699707, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.029690749943256378, + "learning_rate": 0.0001929238073542276, + "loss": 3.1350059509277344, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.029659733176231384, + "learning_rate": 0.00019267625096149903, + "loss": 3.1220788955688477, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.028260033577680588, + "learning_rate": 0.00019242878345462227, + "loss": 3.1000189781188965, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.027554461732506752, + "learning_rate": 0.00019218140503297765, + "loss": 3.1044349670410156, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.027929967269301414, + "learning_rate": 0.00019193411589587332, + "loss": 3.1353745460510254, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.029309319332242012, + "learning_rate": 0.00019168691624254564, + "loss": 3.1578292846679688, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.029183287173509598, + "learning_rate": 0.00019143980627215915, + "loss": 3.1217150688171387, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.02707512304186821, + "learning_rate": 0.00019119278618380555, + "loss": 3.119617223739624, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.03006536327302456, + "learning_rate": 0.00019094585617650476, + "loss": 3.1481480598449707, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.03026784397661686, + "learning_rate": 0.00019069901644920366, + "loss": 3.142770767211914, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.02827967144548893, + "learning_rate": 0.00019045226720077667, + "loss": 3.1582541465759277, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.028892774134874344, + "learning_rate": 0.0001902056086300251, + "loss": 3.1530709266662598, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.03023783676326275, + "learning_rate": 0.0001899590409356773, + "loss": 3.1102521419525146, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.027951840311288834, + "learning_rate": 0.00018971256431638854, + "loss": 3.139113426208496, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.02807226963341236, + "learning_rate": 0.00018946617897074041, + "loss": 3.149428367614746, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.029721451923251152, + "learning_rate": 0.0001892198850972413, + "loss": 3.120518207550049, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.02867680788040161, + "learning_rate": 0.00018897368289432553, + "loss": 3.1109931468963623, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.02663845382630825, + "learning_rate": 0.00018872757256035403, + "loss": 3.116053581237793, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.028366081416606903, + "learning_rate": 0.00018848155429361323, + "loss": 3.1444268226623535, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.028931576758623123, + "learning_rate": 0.00018823562829231567, + "loss": 3.163870096206665, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.02664567157626152, + "learning_rate": 0.0001879897947545996, + "loss": 3.1349146366119385, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.028031354770064354, + "learning_rate": 0.00018774405387852848, + "loss": 3.1638219356536865, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.029421644285321236, + "learning_rate": 0.00018749840586209144, + "loss": 3.1601340770721436, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.027983693405985832, + "learning_rate": 0.00018725285090320245, + "loss": 3.1636788845062256, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.02719871513545513, + "learning_rate": 0.00018700738919970103, + "loss": 3.1246869564056396, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.03030397929251194, + "learning_rate": 0.00018676202094935085, + "loss": 3.1749579906463623, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.029340893030166626, + "learning_rate": 0.000186516746349841, + "loss": 3.1751790046691895, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.027624106034636497, + "learning_rate": 0.0001862715655987847, + "loss": 3.178704023361206, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.02830546908080578, + "learning_rate": 0.00018602647889371957, + "loss": 3.177117347717285, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.03168068826198578, + "learning_rate": 0.00018578148643210764, + "loss": 3.125427722930908, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.033327918499708176, + "learning_rate": 0.00018553658841133483, + "loss": 3.13071870803833, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.218388080596924, + "eval_runtime": 57.611, + "eval_samples_per_second": 42.388, + "eval_steps_per_second": 1.337, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.03336749225854874, + "learning_rate": 0.00018529178502871117, + "loss": 3.064338445663452, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.038479577749967575, + "learning_rate": 0.0001850470764814702, + "loss": 3.0478687286376953, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.032562077045440674, + "learning_rate": 0.0001848024629667691, + "loss": 3.019803762435913, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.0331728458404541, + "learning_rate": 0.00018455794468168887, + "loss": 3.0599141120910645, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.03361169621348381, + "learning_rate": 0.00018431352182323322, + "loss": 3.0315823554992676, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.032973092049360275, + "learning_rate": 0.0001840691945883293, + "loss": 3.025928258895874, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.030388766899704933, + "learning_rate": 0.00018382496317382715, + "loss": 3.070671558380127, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.030899515375494957, + "learning_rate": 0.0001835808277764997, + "loss": 3.012938976287842, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.0335412323474884, + "learning_rate": 0.00018333678859304222, + "loss": 3.0156331062316895, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.035731345415115356, + "learning_rate": 0.00018309284582007274, + "loss": 3.0578935146331787, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.03379053995013237, + "learning_rate": 0.0001828489996541318, + "loss": 3.048701286315918, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.031242679804563522, + "learning_rate": 0.00018260525029168147, + "loss": 3.0832977294921875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.030997611582279205, + "learning_rate": 0.00018236159792910648, + "loss": 3.0663583278656006, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.03397331386804581, + "learning_rate": 0.00018211804276271304, + "loss": 3.0115303993225098, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.03349050134420395, + "learning_rate": 0.00018187458498872914, + "loss": 3.02864933013916, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.031079065054655075, + "learning_rate": 0.00018163122480330433, + "loss": 3.053401470184326, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.029346197843551636, + "learning_rate": 0.00018138796240250955, + "loss": 3.0700926780700684, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.030784208327531815, + "learning_rate": 0.00018114479798233686, + "loss": 3.066525936126709, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.03368140384554863, + "learning_rate": 0.00018090173173869939, + "loss": 3.0956242084503174, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.03356792405247688, + "learning_rate": 0.00018065876386743143, + "loss": 3.0394320487976074, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.029909325763583183, + "learning_rate": 0.00018041589456428754, + "loss": 3.031525135040283, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.030711578205227852, + "learning_rate": 0.0001801731240249434, + "loss": 3.075474977493286, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.03452072665095329, + "learning_rate": 0.00017993045244499463, + "loss": 3.0228147506713867, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.03235075995326042, + "learning_rate": 0.00017968788001995742, + "loss": 3.0664501190185547, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.029671598225831985, + "learning_rate": 0.000179445406945268, + "loss": 3.055302143096924, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.033868469297885895, + "learning_rate": 0.0001792030334162825, + "loss": 3.053016185760498, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.0329158715903759, + "learning_rate": 0.00017896075962827706, + "loss": 3.1017332077026367, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.031983695924282074, + "learning_rate": 0.00017871858577644712, + "loss": 3.0706472396850586, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.02963697351515293, + "learning_rate": 0.0001784765120559078, + "loss": 3.0711426734924316, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.03162391856312752, + "learning_rate": 0.00017823453866169355, + "loss": 3.044867515563965, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.030773654580116272, + "learning_rate": 0.00017799266578875808, + "loss": 3.0963196754455566, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.031220417469739914, + "learning_rate": 0.00017775089363197371, + "loss": 3.0892245769500732, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.030928561463952065, + "learning_rate": 0.00017750922238613198, + "loss": 3.1181323528289795, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.03192073106765747, + "learning_rate": 0.00017726765224594342, + "loss": 3.066584348678589, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.03075495734810829, + "learning_rate": 0.0001770261834060363, + "loss": 3.0436511039733887, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.02895325981080532, + "learning_rate": 0.00017678481606095786, + "loss": 3.123373031616211, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.029738714918494225, + "learning_rate": 0.0001765435504051734, + "loss": 3.1049060821533203, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.031937118619680405, + "learning_rate": 0.0001763023866330664, + "loss": 3.099111318588257, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.029886579141020775, + "learning_rate": 0.00017606132493893793, + "loss": 3.0970301628112793, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.029569782316684723, + "learning_rate": 0.00017582036551700724, + "loss": 3.062499761581421, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.02924283780157566, + "learning_rate": 0.00017557950856141085, + "loss": 3.0618538856506348, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.030955800786614418, + "learning_rate": 0.00017533875426620286, + "loss": 3.0745606422424316, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.03145027533173561, + "learning_rate": 0.00017509810282535497, + "loss": 3.122938632965088, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.028926467522978783, + "learning_rate": 0.00017485755443275513, + "loss": 3.0808773040771484, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.029592076316475868, + "learning_rate": 0.0001746171092822094, + "loss": 3.1167049407958984, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.028070243075489998, + "learning_rate": 0.00017437676756743974, + "loss": 3.0781636238098145, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.028642063960433006, + "learning_rate": 0.00017413652948208524, + "loss": 3.093937397003174, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.0290757454931736, + "learning_rate": 0.00017389639521970136, + "loss": 3.1079559326171875, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.027867283672094345, + "learning_rate": 0.00017365636497376004, + "loss": 3.0885753631591797, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.02896495908498764, + "learning_rate": 0.00017341643893764927, + "loss": 3.0889246463775635, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.02777806483209133, + "learning_rate": 0.000173176617304673, + "loss": 3.084439277648926, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.02904277853667736, + "learning_rate": 0.00017293690026805124, + "loss": 3.125246047973633, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.029010627418756485, + "learning_rate": 0.00017269728802091975, + "loss": 3.111422061920166, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.028178894892334938, + "learning_rate": 0.00017245778075632975, + "loss": 3.115067481994629, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.02763799950480461, + "learning_rate": 0.00017221837866724792, + "loss": 3.1228506565093994, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.028052521869540215, + "learning_rate": 0.00017197908194655615, + "loss": 3.1150546073913574, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.029438253492116928, + "learning_rate": 0.00017173989078705175, + "loss": 3.0921788215637207, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.028195543214678764, + "learning_rate": 0.0001715008053814463, + "loss": 3.1408450603485107, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.02937602624297142, + "learning_rate": 0.00017126182592236682, + "loss": 3.111612319946289, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.028430236503481865, + "learning_rate": 0.00017102295260235472, + "loss": 3.093441963195801, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.027624569833278656, + "learning_rate": 0.0001707841856138659, + "loss": 3.1184372901916504, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.029797039926052094, + "learning_rate": 0.00017054552514927061, + "loss": 3.1074559688568115, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.028728218749165535, + "learning_rate": 0.00017030697140085334, + "loss": 3.161461114883423, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.02704412117600441, + "learning_rate": 0.0001700685245608126, + "loss": 3.13325834274292, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.028067929670214653, + "learning_rate": 0.00016983018482126047, + "loss": 3.093172550201416, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.028848597779870033, + "learning_rate": 0.00016959195237422305, + "loss": 3.1461825370788574, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.02956785447895527, + "learning_rate": 0.00016935382741164, + "loss": 3.141045570373535, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.028214773163199425, + "learning_rate": 0.00016911581012536424, + "loss": 3.136033773422241, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.029155246913433075, + "learning_rate": 0.0001688779007071622, + "loss": 3.0986382961273193, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.02968468703329563, + "learning_rate": 0.00016864009934871275, + "loss": 3.11630916595459, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.028356682509183884, + "learning_rate": 0.0001684024062416087, + "loss": 3.1134822368621826, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.03301718086004257, + "learning_rate": 0.0001681648215773546, + "loss": 3.118591785430908, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.246211528778076, + "eval_runtime": 57.5155, + "eval_samples_per_second": 42.458, + "eval_steps_per_second": 1.339, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.033637888729572296, + "learning_rate": 0.0001679273455473684, + "loss": 3.0192506313323975, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.04057888314127922, + "learning_rate": 0.0001676899783429802, + "loss": 2.977440595626831, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.03579854592680931, + "learning_rate": 0.00016745272015543242, + "loss": 2.9902219772338867, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.03131850063800812, + "learning_rate": 0.0001672155711758799, + "loss": 2.9988656044006348, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.03262636810541153, + "learning_rate": 0.00016697853159538896, + "loss": 3.037747621536255, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.03260403499007225, + "learning_rate": 0.0001667416016049383, + "loss": 2.987508773803711, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.033165134489536285, + "learning_rate": 0.0001665047813954181, + "loss": 3.0511183738708496, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.03303489834070206, + "learning_rate": 0.00016626807115763023, + "loss": 3.0083041191101074, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.03185006603598595, + "learning_rate": 0.00016603147108228754, + "loss": 2.9966228008270264, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.03135651350021362, + "learning_rate": 0.00016579498136001485, + "loss": 3.036759376525879, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.03216925635933876, + "learning_rate": 0.00016555860218134738, + "loss": 3.0304150581359863, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.03231671452522278, + "learning_rate": 0.00016532233373673164, + "loss": 3.0300474166870117, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.03263348713517189, + "learning_rate": 0.00016508617621652488, + "loss": 3.0298163890838623, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.032594650983810425, + "learning_rate": 0.00016485012981099488, + "loss": 2.98789119720459, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.03406650573015213, + "learning_rate": 0.0001646141947103201, + "loss": 3.0102760791778564, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.03379780799150467, + "learning_rate": 0.0001643783711045889, + "loss": 3.0564732551574707, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.03172587603330612, + "learning_rate": 0.0001641426591838005, + "loss": 3.0410943031311035, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.02938298135995865, + "learning_rate": 0.00016390705913786344, + "loss": 3.0517544746398926, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.03206252679228783, + "learning_rate": 0.00016367157115659646, + "loss": 3.0352776050567627, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.033386752009391785, + "learning_rate": 0.00016343619542972808, + "loss": 3.020535945892334, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.033984702080488205, + "learning_rate": 0.0001632009321468962, + "loss": 3.0552730560302734, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.0329931266605854, + "learning_rate": 0.00016296578149764833, + "loss": 3.0075573921203613, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.032101184129714966, + "learning_rate": 0.00016273074367144087, + "loss": 3.046388864517212, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.032873060554265976, + "learning_rate": 0.00016249581885763968, + "loss": 3.0340287685394287, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.030672110617160797, + "learning_rate": 0.0001622610072455194, + "loss": 3.0347838401794434, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.030014054849743843, + "learning_rate": 0.00016202630902426355, + "loss": 3.0315332412719727, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.02932834066450596, + "learning_rate": 0.00016179172438296423, + "loss": 3.084340810775757, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.029181934893131256, + "learning_rate": 0.00016155725351062205, + "loss": 3.032886028289795, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.02966221794486046, + "learning_rate": 0.00016132289659614605, + "loss": 3.032766819000244, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.029713116586208344, + "learning_rate": 0.00016108865382835318, + "loss": 3.028778076171875, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.029243769124150276, + "learning_rate": 0.00016085452539596864, + "loss": 3.0492146015167236, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.02997511625289917, + "learning_rate": 0.0001606205114876256, + "loss": 3.0772786140441895, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.02930292673408985, + "learning_rate": 0.00016038661229186472, + "loss": 3.055750608444214, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.03037521243095398, + "learning_rate": 0.00016015282799713453, + "loss": 3.033710479736328, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.029957240447402, + "learning_rate": 0.0001599191587917905, + "loss": 3.0394721031188965, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.02925589308142662, + "learning_rate": 0.0001596856048640961, + "loss": 3.0878353118896484, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.03003518097102642, + "learning_rate": 0.00015945216640222116, + "loss": 3.048628330230713, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.031545430421829224, + "learning_rate": 0.000159218843594243, + "loss": 3.0775623321533203, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.031310439109802246, + "learning_rate": 0.0001589856366281455, + "loss": 3.01645565032959, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.03074057213962078, + "learning_rate": 0.0001587525456918194, + "loss": 3.0651440620422363, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.028704499825835228, + "learning_rate": 0.0001585195709730619, + "loss": 3.0527148246765137, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.029061976820230484, + "learning_rate": 0.00015828671265957623, + "loss": 3.0854501724243164, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.03034680150449276, + "learning_rate": 0.00015805397093897255, + "loss": 3.056356906890869, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.028715113177895546, + "learning_rate": 0.00015782134599876633, + "loss": 3.0874247550964355, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.030256496742367744, + "learning_rate": 0.00015758883802637953, + "loss": 3.052424430847168, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.02949833683669567, + "learning_rate": 0.00015735644720913923, + "loss": 3.0663650035858154, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.028980428352952003, + "learning_rate": 0.00015712417373427897, + "loss": 3.0700113773345947, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.028526026755571365, + "learning_rate": 0.00015689201778893713, + "loss": 3.0856711864471436, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.029572727158665657, + "learning_rate": 0.00015665997956015743, + "loss": 3.0862035751342773, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.02955584041774273, + "learning_rate": 0.00015642805923488895, + "loss": 3.030802011489868, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.02881692908704281, + "learning_rate": 0.00015619625699998585, + "loss": 3.074392318725586, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.029564999043941498, + "learning_rate": 0.00015596457304220692, + "loss": 3.057894229888916, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.030315004289150238, + "learning_rate": 0.00015573300754821547, + "loss": 3.096233367919922, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.030666006729006767, + "learning_rate": 0.00015550156070458002, + "loss": 3.040799617767334, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.028901133686304092, + "learning_rate": 0.0001552702326977731, + "loss": 3.055051803588867, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.030359387397766113, + "learning_rate": 0.00015503902371417122, + "loss": 3.085972309112549, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.03117106668651104, + "learning_rate": 0.00015480793394005547, + "loss": 3.0786352157592773, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.029238386079669, + "learning_rate": 0.0001545769635616107, + "loss": 3.0375771522521973, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.029324093833565712, + "learning_rate": 0.0001543461127649256, + "loss": 3.1140198707580566, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.02875387854874134, + "learning_rate": 0.00015411538173599231, + "loss": 3.090881109237671, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.029023943468928337, + "learning_rate": 0.00015388477066070674, + "loss": 3.0755701065063477, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.0291463453322649, + "learning_rate": 0.00015365427972486827, + "loss": 3.0723776817321777, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.02795177884399891, + "learning_rate": 0.00015342390911417904, + "loss": 3.058993339538574, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.029499078169465065, + "learning_rate": 0.00015319365901424452, + "loss": 3.0856404304504395, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.029889609664678574, + "learning_rate": 0.0001529635296105732, + "loss": 3.0622942447662354, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.03107610158622265, + "learning_rate": 0.0001527335210885761, + "loss": 3.114957332611084, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.02967257983982563, + "learning_rate": 0.00015250363363356685, + "loss": 3.088871479034424, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.029068559408187866, + "learning_rate": 0.00015227386743076155, + "loss": 3.1103081703186035, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.029334546998143196, + "learning_rate": 0.00015204422266527911, + "loss": 3.0841622352600098, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.029259610921144485, + "learning_rate": 0.00015181469952213968, + "loss": 3.0662848949432373, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.030072513967752457, + "learning_rate": 0.0001515852981862663, + "loss": 3.0900328159332275, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.033321622759103775, + "learning_rate": 0.00015135601884248303, + "loss": 3.0871129035949707, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.271463394165039, + "eval_runtime": 57.9048, + "eval_samples_per_second": 42.173, + "eval_steps_per_second": 1.33, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.03635408729314804, + "learning_rate": 0.0001511268616755166, + "loss": 2.9576284885406494, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.03927835449576378, + "learning_rate": 0.00015089782686999445, + "loss": 2.999624729156494, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.03371119126677513, + "learning_rate": 0.00015066891461044588, + "loss": 2.971052646636963, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.033049341291189194, + "learning_rate": 0.00015044012508130138, + "loss": 2.980811595916748, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.03830729424953461, + "learning_rate": 0.00015021145846689254, + "loss": 2.9985947608947754, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.036573827266693115, + "learning_rate": 0.00014998291495145206, + "loss": 2.9892568588256836, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.031196987256407738, + "learning_rate": 0.00014975449471911294, + "loss": 2.9854254722595215, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.03376495838165283, + "learning_rate": 0.0001495261979539097, + "loss": 2.9840264320373535, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.036272983998060226, + "learning_rate": 0.0001492980248397766, + "loss": 3.0168871879577637, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.03328242897987366, + "learning_rate": 0.00014906997556054865, + "loss": 2.9815053939819336, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.03158426657319069, + "learning_rate": 0.00014884205029996113, + "loss": 2.994525909423828, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.03320910036563873, + "learning_rate": 0.00014861424924164925, + "loss": 3.008638858795166, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.03160254284739494, + "learning_rate": 0.00014838657256914832, + "loss": 2.979243040084839, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.03158991411328316, + "learning_rate": 0.0001481590204658931, + "loss": 2.938634157180786, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.031116485595703125, + "learning_rate": 0.00014793159311521833, + "loss": 2.9957656860351562, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.030757686123251915, + "learning_rate": 0.00014770429070035812, + "loss": 3.0106635093688965, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.02992827817797661, + "learning_rate": 0.00014747711340444601, + "loss": 2.996793746948242, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.0330309197306633, + "learning_rate": 0.00014725006141051434, + "loss": 2.9859161376953125, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.029488038271665573, + "learning_rate": 0.00014702313490149512, + "loss": 3.0081536769866943, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.03095114603638649, + "learning_rate": 0.00014679633406021892, + "loss": 3.006385087966919, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.03194110840559006, + "learning_rate": 0.00014656965906941494, + "loss": 3.0125250816345215, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.03306785598397255, + "learning_rate": 0.00014634311011171112, + "loss": 2.9640283584594727, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.030883541330695152, + "learning_rate": 0.00014611668736963396, + "loss": 3.0223653316497803, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.03224025294184685, + "learning_rate": 0.00014589039102560822, + "loss": 3.0274910926818848, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.03182433918118477, + "learning_rate": 0.00014566422126195656, + "loss": 3.047736406326294, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.03387221321463585, + "learning_rate": 0.00014543817826089992, + "loss": 2.9821343421936035, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.03407924994826317, + "learning_rate": 0.0001452122622045574, + "loss": 3.041207790374756, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.031546324491500854, + "learning_rate": 0.00014498647327494506, + "loss": 3.0298609733581543, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.03354770317673683, + "learning_rate": 0.00014476081165397715, + "loss": 3.0212411880493164, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.031692177057266235, + "learning_rate": 0.00014453527752346517, + "loss": 3.007350444793701, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.03218817710876465, + "learning_rate": 0.00014430987106511795, + "loss": 3.0019829273223877, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.03233771771192551, + "learning_rate": 0.00014408459246054126, + "loss": 3.0400614738464355, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.028998777270317078, + "learning_rate": 0.00014385944189123794, + "loss": 3.022221088409424, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.030422436073422432, + "learning_rate": 0.00014363441953860805, + "loss": 3.0071797370910645, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.030207227915525436, + "learning_rate": 0.00014340952558394778, + "loss": 2.9762964248657227, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.0305719543248415, + "learning_rate": 0.00014318476020845018, + "loss": 3.075176239013672, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.03173545375466347, + "learning_rate": 0.0001429601235932047, + "loss": 3.028982162475586, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.02799576334655285, + "learning_rate": 0.00014273561591919705, + "loss": 3.0242385864257812, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.03140150383114815, + "learning_rate": 0.00014251123736730884, + "loss": 3.012221336364746, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.029416466131806374, + "learning_rate": 0.00014228698811831787, + "loss": 3.0037949085235596, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.028385788202285767, + "learning_rate": 0.0001420628683528978, + "loss": 3.0350992679595947, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.031597770750522614, + "learning_rate": 0.00014183887825161779, + "loss": 3.065139055252075, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.02926146425306797, + "learning_rate": 0.0001416150179949428, + "loss": 3.014230728149414, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.02970231883227825, + "learning_rate": 0.0001413912877632326, + "loss": 3.06636381149292, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.029684338718652725, + "learning_rate": 0.00014116768773674306, + "loss": 3.024610996246338, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.02908361330628395, + "learning_rate": 0.00014094421809562438, + "loss": 3.0093719959259033, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.029891187325119972, + "learning_rate": 0.00014072087901992209, + "loss": 3.0735411643981934, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.028607673943042755, + "learning_rate": 0.00014049767068957645, + "loss": 3.044274091720581, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.029806701466441154, + "learning_rate": 0.00014027459328442242, + "loss": 3.0449328422546387, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.03161340951919556, + "learning_rate": 0.00014005164698418948, + "loss": 3.046813488006592, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.028860243037343025, + "learning_rate": 0.0001398288319685013, + "loss": 3.011828660964966, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.0285879448056221, + "learning_rate": 0.00013960614841687595, + "loss": 3.043640613555908, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.0296401996165514, + "learning_rate": 0.00013938359650872558, + "loss": 3.071589469909668, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.02807026542723179, + "learning_rate": 0.00013916117642335623, + "loss": 3.018320083618164, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.029037052765488625, + "learning_rate": 0.00013893888833996774, + "loss": 3.0471763610839844, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.029226379469037056, + "learning_rate": 0.00013871673243765362, + "loss": 3.037463426589966, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.029322804883122444, + "learning_rate": 0.00013849470889540098, + "loss": 3.0472755432128906, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.028253905475139618, + "learning_rate": 0.00013827281789208996, + "loss": 3.0265414714813232, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.028875533491373062, + "learning_rate": 0.00013805105960649423, + "loss": 3.063941717147827, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.028460513800382614, + "learning_rate": 0.00013782943421728042, + "loss": 3.0512683391571045, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.02951696142554283, + "learning_rate": 0.00013760794190300827, + "loss": 3.0528573989868164, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.030593490228056908, + "learning_rate": 0.00013738658284212973, + "loss": 3.057919979095459, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.028074532747268677, + "learning_rate": 0.00013716535721299016, + "loss": 3.0263280868530273, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.02955678664147854, + "learning_rate": 0.0001369442651938271, + "loss": 3.081322193145752, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.029724644497036934, + "learning_rate": 0.00013672330696277014, + "loss": 3.059583902359009, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.029713226482272148, + "learning_rate": 0.00013650248269784143, + "loss": 3.018678903579712, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.029758386313915253, + "learning_rate": 0.00013628179257695508, + "loss": 3.0379555225372314, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.028971998021006584, + "learning_rate": 0.00013606123677791727, + "loss": 3.041267156600952, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.02968725562095642, + "learning_rate": 0.00013584081547842565, + "loss": 3.0675506591796875, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.029876509681344032, + "learning_rate": 0.00013562052885606957, + "loss": 3.051586151123047, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.02853119559586048, + "learning_rate": 0.00013540037708833043, + "loss": 3.0335168838500977, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.03374241292476654, + "learning_rate": 0.00013518036035258012, + "loss": 3.027909278869629, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.030695006251335144, + "learning_rate": 0.0001349604788260823, + "loss": 2.9685583114624023, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0339728482067585, + "learning_rate": 0.00013474073268599152, + "loss": 2.936551570892334, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.03372679278254509, + "learning_rate": 0.00013452112210935339, + "loss": 2.9631872177124023, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.03172554448246956, + "learning_rate": 0.00013430164727310399, + "loss": 2.969510078430176, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.030516354367136955, + "learning_rate": 0.00013408230835407027, + "loss": 2.930342674255371, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.03261839970946312, + "learning_rate": 0.00013386310552896977, + "loss": 2.9920477867126465, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.030618635937571526, + "learning_rate": 0.00013364403897441016, + "loss": 2.9693596363067627, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.030576011165976524, + "learning_rate": 0.00013342510886688944, + "loss": 2.929804801940918, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.03224785998463631, + "learning_rate": 0.00013320631538279548, + "loss": 2.95280122756958, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.030605586245656013, + "learning_rate": 0.00013298765869840658, + "loss": 2.931358575820923, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.031206639483571053, + "learning_rate": 0.00013276913898989013, + "loss": 2.962998151779175, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.031317099928855896, + "learning_rate": 0.00013255075643330369, + "loss": 2.9706358909606934, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.031614501029253006, + "learning_rate": 0.00013233251120459407, + "loss": 2.9326977729797363, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.031002728268504143, + "learning_rate": 0.0001321144034795976, + "loss": 2.99625301361084, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.03216253221035004, + "learning_rate": 0.00013189643343403977, + "loss": 2.920912027359009, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.033380527049303055, + "learning_rate": 0.00013167860124353476, + "loss": 2.9688234329223633, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.03085758350789547, + "learning_rate": 0.00013146090708358657, + "loss": 2.9811222553253174, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.03258587792515755, + "learning_rate": 0.00013124335112958704, + "loss": 2.932366371154785, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.031639691442251205, + "learning_rate": 0.0001310259335568172, + "loss": 2.988865613937378, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.03144839406013489, + "learning_rate": 0.00013080865454044647, + "loss": 2.9750723838806152, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.030864672735333443, + "learning_rate": 0.00013059151425553263, + "loss": 2.999329090118408, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.030679022893309593, + "learning_rate": 0.00013037451287702183, + "loss": 2.974264144897461, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.03236627206206322, + "learning_rate": 0.0001301576505797479, + "loss": 2.9808404445648193, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.03115275874733925, + "learning_rate": 0.00012994092753843293, + "loss": 3.0216290950775146, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.029868317767977715, + "learning_rate": 0.00012972434392768687, + "loss": 2.9739654064178467, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.03446527197957039, + "learning_rate": 0.00012950789992200714, + "loss": 2.9565515518188477, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.033536117523908615, + "learning_rate": 0.00012929159569577886, + "loss": 2.958047389984131, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.03113192319869995, + "learning_rate": 0.0001290754314232743, + "loss": 2.9662246704101562, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.03089931420981884, + "learning_rate": 0.00012885940727865334, + "loss": 2.9972119331359863, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.030822310596704483, + "learning_rate": 0.00012864352343596247, + "loss": 3.020500421524048, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.032264843583106995, + "learning_rate": 0.0001284277800691355, + "loss": 2.9895882606506348, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.031226349994540215, + "learning_rate": 0.00012821217735199298, + "loss": 2.958217144012451, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.03110436722636223, + "learning_rate": 0.00012799671545824212, + "loss": 2.9884791374206543, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.03358910232782364, + "learning_rate": 0.0001277813945614768, + "loss": 2.9908103942871094, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.03348501771688461, + "learning_rate": 0.00012756621483517682, + "loss": 2.985677480697632, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.029595404863357544, + "learning_rate": 0.00012735117645270905, + "loss": 2.979318380355835, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.031459368765354156, + "learning_rate": 0.00012713627958732567, + "loss": 2.9921798706054688, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.03067564032971859, + "learning_rate": 0.00012692152441216539, + "loss": 2.977778434753418, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.030506785959005356, + "learning_rate": 0.0001267069111002525, + "loss": 2.9841768741607666, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.029978664591908455, + "learning_rate": 0.00012649243982449718, + "loss": 2.995443344116211, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.030900927260518074, + "learning_rate": 0.00012627811075769502, + "loss": 3.0177016258239746, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.029710259288549423, + "learning_rate": 0.00012606392407252687, + "loss": 3.01532244682312, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.031190430745482445, + "learning_rate": 0.00012584987994155943, + "loss": 2.9791834354400635, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.03150483965873718, + "learning_rate": 0.00012563597853724388, + "loss": 2.998800754547119, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.030249621719121933, + "learning_rate": 0.000125422220031917, + "loss": 3.0187838077545166, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.031052686274051666, + "learning_rate": 0.00012520860459779972, + "loss": 2.984062433242798, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.02961024083197117, + "learning_rate": 0.0001249951324069986, + "loss": 2.970818042755127, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.03020382858812809, + "learning_rate": 0.0001247818036315042, + "loss": 2.9913196563720703, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.031914591789245605, + "learning_rate": 0.00012456861844319155, + "loss": 2.993145704269409, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.03061537817120552, + "learning_rate": 0.00012435557701382012, + "loss": 3.0131571292877197, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.029960963875055313, + "learning_rate": 0.0001241426795150336, + "loss": 3.0017619132995605, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.030348235741257668, + "learning_rate": 0.00012392992611835973, + "loss": 3.0220556259155273, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.03030729480087757, + "learning_rate": 0.0001237173169952098, + "loss": 3.0499038696289062, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.030809229239821434, + "learning_rate": 0.00012350485231687954, + "loss": 2.986651659011841, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.029438305646181107, + "learning_rate": 0.0001232925322545476, + "loss": 3.0351314544677734, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.029887216165661812, + "learning_rate": 0.0001230803569792765, + "loss": 3.021660327911377, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.02955588512122631, + "learning_rate": 0.0001228683266620121, + "loss": 2.9990694522857666, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.029586469754576683, + "learning_rate": 0.00012265644147358326, + "loss": 3.0296952724456787, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.030632557347416878, + "learning_rate": 0.00012244470158470226, + "loss": 3.0124430656433105, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.030579863116145134, + "learning_rate": 0.00012223310716596387, + "loss": 3.0381762981414795, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.03282996267080307, + "learning_rate": 0.00012202165838784601, + "loss": 3.0329041481018066, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.031176412478089333, + "learning_rate": 0.00012181035542070913, + "loss": 3.037598133087158, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.028640294447541237, + "learning_rate": 0.00012159919843479617, + "loss": 3.0004701614379883, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.0326533280313015, + "learning_rate": 0.00012138818760023257, + "loss": 3.0328853130340576, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.03041139990091324, + "learning_rate": 0.00012117732308702592, + "loss": 3.0149893760681152, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.030366195365786552, + "learning_rate": 0.00012096660506506605, + "loss": 3.0379695892333984, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.03106893040239811, + "learning_rate": 0.00012075603370412443, + "loss": 3.0284576416015625, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.0288887657225132, + "learning_rate": 0.00012054560917385476, + "loss": 3.028569221496582, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.03154266998171806, + "learning_rate": 0.00012033533164379225, + "loss": 3.0737850666046143, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.0305064395070076, + "learning_rate": 0.0001201252012833537, + "loss": 3.0744194984436035, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.030967431142926216, + "learning_rate": 0.00011991521826183747, + "loss": 3.0170607566833496, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.036945272237062454, + "learning_rate": 0.00011970538274842282, + "loss": 3.0355610847473145, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.329805374145508, + "eval_runtime": 64.1339, + "eval_samples_per_second": 38.077, + "eval_steps_per_second": 1.201, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.03236053138971329, + "learning_rate": 0.00011949569491217073, + "loss": 2.9230728149414062, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.0395631417632103, + "learning_rate": 0.00011928615492202269, + "loss": 2.92578125, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.0361955352127552, + "learning_rate": 0.0001190767629468014, + "loss": 2.937826633453369, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.035918429493904114, + "learning_rate": 0.00011886751915521007, + "loss": 2.9481167793273926, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.03309318795800209, + "learning_rate": 0.00011865842371583278, + "loss": 2.8919506072998047, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0327315591275692, + "learning_rate": 0.00011844947679713396, + "loss": 2.903585433959961, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.033141329884529114, + "learning_rate": 0.00011824067856745812, + "loss": 2.942802667617798, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.033966805785894394, + "learning_rate": 0.00011803202919503064, + "loss": 2.9210848808288574, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.03218021243810654, + "learning_rate": 0.00011782352884795615, + "loss": 2.933588981628418, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.03272828459739685, + "learning_rate": 0.00011761517769421983, + "loss": 2.9202046394348145, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.030764421448111534, + "learning_rate": 0.00011740697590168635, + "loss": 2.916121006011963, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.031897611916065216, + "learning_rate": 0.00011719892363810018, + "loss": 2.957679033279419, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.03199796751141548, + "learning_rate": 0.00011699102107108539, + "loss": 2.9276621341705322, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.030233927071094513, + "learning_rate": 0.00011678326836814507, + "loss": 2.9006078243255615, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.03035062551498413, + "learning_rate": 0.00011657566569666198, + "loss": 2.920341730117798, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.03029404580593109, + "learning_rate": 0.00011636821322389777, + "loss": 2.937875747680664, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.029371997341513634, + "learning_rate": 0.00011616091111699333, + "loss": 2.9832730293273926, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.030163884162902832, + "learning_rate": 0.00011595375954296786, + "loss": 2.94118332862854, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.03072655200958252, + "learning_rate": 0.00011574675866871997, + "loss": 2.9632580280303955, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.02975962497293949, + "learning_rate": 0.00011553990866102651, + "loss": 2.9346470832824707, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.029814405366778374, + "learning_rate": 0.00011533320968654265, + "loss": 2.9451403617858887, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.030135225504636765, + "learning_rate": 0.00011512666191180204, + "loss": 2.944918155670166, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.029395751655101776, + "learning_rate": 0.00011492026550321653, + "loss": 2.925586223602295, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.0306396484375, + "learning_rate": 0.00011471402062707607, + "loss": 2.9487359523773193, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.03110465221107006, + "learning_rate": 0.00011450792744954827, + "loss": 2.963491678237915, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.03155078738927841, + "learning_rate": 0.00011430198613667861, + "loss": 2.9355576038360596, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.030103784054517746, + "learning_rate": 0.00011409619685439064, + "loss": 2.953045606613159, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.030827803537249565, + "learning_rate": 0.00011389055976848477, + "loss": 2.968574047088623, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.0310911163687706, + "learning_rate": 0.00011368507504463914, + "loss": 2.977902412414551, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.030260782688856125, + "learning_rate": 0.00011347974284840914, + "loss": 2.94873309135437, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.030654428526759148, + "learning_rate": 0.00011327456334522725, + "loss": 2.9282753467559814, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.02993479184806347, + "learning_rate": 0.00011306953670040275, + "loss": 2.97076416015625, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.030244970694184303, + "learning_rate": 0.00011286466307912184, + "loss": 2.960441827774048, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.030003050342202187, + "learning_rate": 0.00011265994264644783, + "loss": 2.9672164916992188, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.030057862401008606, + "learning_rate": 0.00011245537556732001, + "loss": 2.961918830871582, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.029379919171333313, + "learning_rate": 0.00011225096200655442, + "loss": 2.9918394088745117, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.03060895763337612, + "learning_rate": 0.00011204670212884342, + "loss": 2.94891619682312, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.030228130519390106, + "learning_rate": 0.00011184259609875557, + "loss": 2.9726617336273193, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.029833678156137466, + "learning_rate": 0.00011163864408073522, + "loss": 2.9494335651397705, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.030842263251543045, + "learning_rate": 0.00011143484623910293, + "loss": 2.9580483436584473, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.029707176610827446, + "learning_rate": 0.00011123120273805496, + "loss": 2.963684320449829, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.030966786667704582, + "learning_rate": 0.0001110277137416632, + "loss": 2.9710958003997803, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.03137032687664032, + "learning_rate": 0.00011082437941387512, + "loss": 3.003589391708374, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.03123190440237522, + "learning_rate": 0.00011062119991851322, + "loss": 2.9742860794067383, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.030608588829636574, + "learning_rate": 0.000110418175419276, + "loss": 2.97404408454895, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.030018644407391548, + "learning_rate": 0.00011021530607973626, + "loss": 2.972231864929199, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.03056846372783184, + "learning_rate": 0.00011001259206334235, + "loss": 2.9871716499328613, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.02888435684144497, + "learning_rate": 0.00010981003353341721, + "loss": 2.9774169921875, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.03173990175127983, + "learning_rate": 0.00010960763065315864, + "loss": 2.941751480102539, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.030047327280044556, + "learning_rate": 0.00010940538358563906, + "loss": 2.990516185760498, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.030729815363883972, + "learning_rate": 0.00010920329249380509, + "loss": 2.9745054244995117, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.030600983649492264, + "learning_rate": 0.00010900135754047799, + "loss": 2.9676642417907715, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.03208782523870468, + "learning_rate": 0.00010879957888835307, + "loss": 3.006941556930542, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.031162556260824203, + "learning_rate": 0.00010859795669999981, + "loss": 2.9740395545959473, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.031870320439338684, + "learning_rate": 0.00010839649113786137, + "loss": 2.9809296131134033, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.03184300288558006, + "learning_rate": 0.00010819518236425514, + "loss": 2.9781875610351562, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.029629364609718323, + "learning_rate": 0.00010799403054137197, + "loss": 2.9912891387939453, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.03322628512978554, + "learning_rate": 0.00010779303583127609, + "loss": 2.9957847595214844, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.0322413295507431, + "learning_rate": 0.00010759219839590535, + "loss": 2.9894957542419434, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.030753307044506073, + "learning_rate": 0.00010739151839707089, + "loss": 3.0320589542388916, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.03447900339961052, + "learning_rate": 0.00010719099599645706, + "loss": 3.0189106464385986, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.032016605138778687, + "learning_rate": 0.0001069906313556208, + "loss": 3.0162312984466553, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.030838409438729286, + "learning_rate": 0.00010679042463599263, + "loss": 3.000263214111328, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.034556787461042404, + "learning_rate": 0.00010659037599887545, + "loss": 2.996680736541748, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.03266173601150513, + "learning_rate": 0.0001063904856054446, + "loss": 2.990510940551758, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.030973197892308235, + "learning_rate": 0.00010619075361674836, + "loss": 2.9767255783081055, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.03442178666591644, + "learning_rate": 0.0001059911801937071, + "loss": 2.9941515922546387, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.03215927258133888, + "learning_rate": 0.00010579176549711365, + "loss": 3.006618022918701, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.030949831008911133, + "learning_rate": 0.00010559250968763265, + "loss": 3.004136085510254, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.03256693854928017, + "learning_rate": 0.00010539341292580086, + "loss": 3.0229580402374268, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.02986278384923935, + "learning_rate": 0.00010519447537202729, + "loss": 2.9580869674682617, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.03597614914178848, + "learning_rate": 0.00010499569718659201, + "loss": 2.995757579803467, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.359405517578125, + "eval_runtime": 68.1721, + "eval_samples_per_second": 35.821, + "eval_steps_per_second": 1.129, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.03432514891028404, + "learning_rate": 0.00010479707852964713, + "loss": 2.922243118286133, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.0345693901181221, + "learning_rate": 0.00010459861956121611, + "loss": 2.9114043712615967, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.0349399633705616, + "learning_rate": 0.00010440032044119383, + "loss": 2.8668057918548584, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.03161955252289772, + "learning_rate": 0.00010420218132934614, + "loss": 2.8996763229370117, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.032124556601047516, + "learning_rate": 0.00010400420238531023, + "loss": 2.8662075996398926, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.03291172534227371, + "learning_rate": 0.00010380638376859416, + "loss": 2.9088220596313477, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.033552397042512894, + "learning_rate": 0.00010360872563857682, + "loss": 2.9076967239379883, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.03218761831521988, + "learning_rate": 0.0001034112281545079, + "loss": 2.8726117610931396, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.03289451450109482, + "learning_rate": 0.00010321389147550725, + "loss": 2.9127345085144043, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.033019524067640305, + "learning_rate": 0.00010301671576056588, + "loss": 2.920897960662842, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.03153533488512039, + "learning_rate": 0.00010281970116854436, + "loss": 2.9124374389648438, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.032293178141117096, + "learning_rate": 0.00010262284785817392, + "loss": 2.877277374267578, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.031969670206308365, + "learning_rate": 0.00010242615598805574, + "loss": 2.921635150909424, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.03023488074541092, + "learning_rate": 0.00010222962571666088, + "loss": 2.9336061477661133, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.03215106576681137, + "learning_rate": 0.00010203325720233032, + "loss": 2.905877113342285, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.030415233224630356, + "learning_rate": 0.00010183705060327433, + "loss": 2.909331798553467, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.031150560826063156, + "learning_rate": 0.00010164100607757346, + "loss": 2.9302995204925537, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.03152180463075638, + "learning_rate": 0.00010144512378317687, + "loss": 2.9171547889709473, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.02996763028204441, + "learning_rate": 0.00010124940387790354, + "loss": 2.916459798812866, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.030544105917215347, + "learning_rate": 0.00010105384651944148, + "loss": 2.898505449295044, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.030963370576500893, + "learning_rate": 0.00010085845186534769, + "loss": 2.948704719543457, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.03191790357232094, + "learning_rate": 0.00010066322007304819, + "loss": 2.9182960987091064, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.02956460975110531, + "learning_rate": 0.00010046815129983757, + "loss": 2.9211082458496094, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.031227655708789825, + "learning_rate": 0.00010027324570287925, + "loss": 2.942269802093506, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.031859587877988815, + "learning_rate": 0.00010007850343920519, + "loss": 2.9169349670410156, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.03045276179909706, + "learning_rate": 9.988392466571572e-05, + "loss": 2.9240684509277344, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.031577471643686295, + "learning_rate": 9.968950953917945e-05, + "loss": 2.9135894775390625, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.031031791120767593, + "learning_rate": 9.949525821623309e-05, + "loss": 2.913355827331543, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.030012110248208046, + "learning_rate": 9.930117085338156e-05, + "loss": 2.8860886096954346, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.029772799462080002, + "learning_rate": 9.910724760699731e-05, + "loss": 2.974332809448242, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.029961293563246727, + "learning_rate": 9.891348863332093e-05, + "loss": 2.943694591522217, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.029537590220570564, + "learning_rate": 9.871989408846051e-05, + "loss": 2.93180513381958, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.030834559351205826, + "learning_rate": 9.852646412839178e-05, + "loss": 2.9442358016967773, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.02998523600399494, + "learning_rate": 9.833319890895756e-05, + "loss": 2.9336910247802734, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.030030962079763412, + "learning_rate": 9.814009858586817e-05, + "loss": 2.93593692779541, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.030250679701566696, + "learning_rate": 9.794716331470131e-05, + "loss": 2.9370553493499756, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.02942339889705181, + "learning_rate": 9.77543932509012e-05, + "loss": 2.9406495094299316, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.030448907986283302, + "learning_rate": 9.756178854977925e-05, + "loss": 2.9552698135375977, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.03078317642211914, + "learning_rate": 9.736934936651362e-05, + "loss": 2.933631420135498, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.029188675805926323, + "learning_rate": 9.717707585614916e-05, + "loss": 2.944002628326416, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.03153637796640396, + "learning_rate": 9.698496817359698e-05, + "loss": 2.9708166122436523, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.030299564823508263, + "learning_rate": 9.679302647363476e-05, + "loss": 2.9570717811584473, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.03169458732008934, + "learning_rate": 9.660125091090675e-05, + "loss": 2.942774772644043, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.031527288258075714, + "learning_rate": 9.64096416399228e-05, + "loss": 2.9711742401123047, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.0314360186457634, + "learning_rate": 9.621819881505918e-05, + "loss": 2.9509239196777344, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.032102152705192566, + "learning_rate": 9.602692259055767e-05, + "loss": 2.9443840980529785, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.03111223131418228, + "learning_rate": 9.583581312052646e-05, + "loss": 2.972594738006592, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.0322272814810276, + "learning_rate": 9.564487055893867e-05, + "loss": 2.9361982345581055, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.033640168607234955, + "learning_rate": 9.545409505963338e-05, + "loss": 2.9726009368896484, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.02979348786175251, + "learning_rate": 9.526348677631499e-05, + "loss": 2.9978041648864746, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.030464721843600273, + "learning_rate": 9.507304586255311e-05, + "loss": 2.949810028076172, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.03170578181743622, + "learning_rate": 9.488277247178267e-05, + "loss": 2.9677181243896484, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.02996704913675785, + "learning_rate": 9.469266675730319e-05, + "loss": 2.974087953567505, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.030277637764811516, + "learning_rate": 9.450272887227983e-05, + "loss": 2.95644211769104, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.030859284102916718, + "learning_rate": 9.431295896974182e-05, + "loss": 2.957258701324463, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.029737796634435654, + "learning_rate": 9.412335720258341e-05, + "loss": 2.9502854347229004, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.029822375625371933, + "learning_rate": 9.393392372356335e-05, + "loss": 2.9625749588012695, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.030455606058239937, + "learning_rate": 9.374465868530477e-05, + "loss": 2.9669153690338135, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.029491793364286423, + "learning_rate": 9.355556224029515e-05, + "loss": 2.952991247177124, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.030326122418045998, + "learning_rate": 9.336663454088593e-05, + "loss": 2.978355646133423, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.03011198714375496, + "learning_rate": 9.317787573929282e-05, + "loss": 2.976996421813965, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.029053322970867157, + "learning_rate": 9.298928598759541e-05, + "loss": 2.9754371643066406, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.02925761416554451, + "learning_rate": 9.280086543773698e-05, + "loss": 2.94271183013916, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.029678862541913986, + "learning_rate": 9.26126142415247e-05, + "loss": 2.9386143684387207, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.031623054295778275, + "learning_rate": 9.2424532550629e-05, + "loss": 2.9660329818725586, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.030133802443742752, + "learning_rate": 9.223662051658408e-05, + "loss": 2.9580249786376953, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.029482144862413406, + "learning_rate": 9.204887829078709e-05, + "loss": 2.944474697113037, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.030708568170666695, + "learning_rate": 9.186130602449861e-05, + "loss": 2.967862367630005, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.028793595731258392, + "learning_rate": 9.167390386884224e-05, + "loss": 2.9581642150878906, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.03083081543445587, + "learning_rate": 9.148667197480455e-05, + "loss": 2.957261323928833, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.030180038884282112, + "learning_rate": 9.129961049323494e-05, + "loss": 2.9579484462738037, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.034881673753261566, + "learning_rate": 9.111271957484519e-05, + "loss": 2.9007070064544678, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.38953161239624, + "eval_runtime": 71.4964, + "eval_samples_per_second": 34.156, + "eval_steps_per_second": 1.077, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.0335528701543808, + "learning_rate": 9.092599937021031e-05, + "loss": 2.8542537689208984, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.03163984417915344, + "learning_rate": 9.073945002976715e-05, + "loss": 2.864593505859375, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.034005917608737946, + "learning_rate": 9.055307170381522e-05, + "loss": 2.855072498321533, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.03235425800085068, + "learning_rate": 9.036686454251613e-05, + "loss": 2.8504371643066406, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.03171464428305626, + "learning_rate": 9.018082869589369e-05, + "loss": 2.8924291133880615, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.0333196297287941, + "learning_rate": 8.999496431383368e-05, + "loss": 2.8764119148254395, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.03050111047923565, + "learning_rate": 8.980927154608341e-05, + "loss": 2.874119281768799, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.032114267349243164, + "learning_rate": 8.962375054225255e-05, + "loss": 2.9059693813323975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.03141588345170021, + "learning_rate": 8.943840145181177e-05, + "loss": 2.879429340362549, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.03201865404844284, + "learning_rate": 8.925322442409361e-05, + "loss": 2.898519277572632, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.03152940049767494, + "learning_rate": 8.906821960829163e-05, + "loss": 2.866131544113159, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.032520636916160583, + "learning_rate": 8.888338715346113e-05, + "loss": 2.888209819793701, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.03244909271597862, + "learning_rate": 8.869872720851831e-05, + "loss": 2.888913631439209, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.03160687908530235, + "learning_rate": 8.851423992224012e-05, + "loss": 2.905120372772217, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.031518299132585526, + "learning_rate": 8.832992544326473e-05, + "loss": 2.8985815048217773, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.0304370429366827, + "learning_rate": 8.814578392009104e-05, + "loss": 2.8531291484832764, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.03239843621850014, + "learning_rate": 8.796181550107857e-05, + "loss": 2.876098394393921, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.03194732964038849, + "learning_rate": 8.777802033444712e-05, + "loss": 2.908271074295044, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.03145617991685867, + "learning_rate": 8.75943985682774e-05, + "loss": 2.874128580093384, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.031182898208498955, + "learning_rate": 8.741095035051017e-05, + "loss": 2.894904613494873, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.0323818176984787, + "learning_rate": 8.722767582894613e-05, + "loss": 2.9005861282348633, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.030969973653554916, + "learning_rate": 8.704457515124636e-05, + "loss": 2.9236338138580322, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.03238993510603905, + "learning_rate": 8.686164846493176e-05, + "loss": 2.8811020851135254, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.030876649543642998, + "learning_rate": 8.667889591738317e-05, + "loss": 2.8802058696746826, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.032251305878162384, + "learning_rate": 8.649631765584083e-05, + "loss": 2.882112503051758, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.03334304690361023, + "learning_rate": 8.631391382740482e-05, + "loss": 2.9290661811828613, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.03299225866794586, + "learning_rate": 8.613168457903459e-05, + "loss": 2.908268928527832, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.03335456922650337, + "learning_rate": 8.594963005754901e-05, + "loss": 2.9156572818756104, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.030786098912358284, + "learning_rate": 8.57677504096261e-05, + "loss": 2.8981709480285645, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.03379470854997635, + "learning_rate": 8.558604578180301e-05, + "loss": 2.9483864307403564, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.03020016849040985, + "learning_rate": 8.540451632047593e-05, + "loss": 2.908958911895752, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.032774876803159714, + "learning_rate": 8.522316217189972e-05, + "loss": 2.899505138397217, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.030589981004595757, + "learning_rate": 8.504198348218821e-05, + "loss": 2.9257779121398926, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.03366618603467941, + "learning_rate": 8.486098039731384e-05, + "loss": 2.8749284744262695, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.03038695827126503, + "learning_rate": 8.46801530631075e-05, + "loss": 2.9127252101898193, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.03400666266679764, + "learning_rate": 8.449950162525859e-05, + "loss": 2.922375202178955, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.031702589243650436, + "learning_rate": 8.431902622931443e-05, + "loss": 2.8886446952819824, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.03223840892314911, + "learning_rate": 8.413872702068119e-05, + "loss": 2.906278610229492, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.0327419750392437, + "learning_rate": 8.395860414462238e-05, + "loss": 2.9436042308807373, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.030698910355567932, + "learning_rate": 8.377865774625985e-05, + "loss": 2.897615909576416, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.03250030800700188, + "learning_rate": 8.359888797057316e-05, + "loss": 2.928028106689453, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.030623450875282288, + "learning_rate": 8.341929496239958e-05, + "loss": 2.9268975257873535, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.030981825664639473, + "learning_rate": 8.323987886643402e-05, + "loss": 2.938185691833496, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.03139727562665939, + "learning_rate": 8.306063982722855e-05, + "loss": 2.9397618770599365, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.029906271025538445, + "learning_rate": 8.288157798919318e-05, + "loss": 2.921111583709717, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.03110392577946186, + "learning_rate": 8.27026934965945e-05, + "loss": 2.899641513824463, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.03060143068432808, + "learning_rate": 8.25239864935566e-05, + "loss": 2.922635078430176, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.03061371110379696, + "learning_rate": 8.23454571240605e-05, + "loss": 2.925189733505249, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.03080526366829872, + "learning_rate": 8.216710553194416e-05, + "loss": 2.90329647064209, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.030412547290325165, + "learning_rate": 8.198893186090222e-05, + "loss": 2.9444844722747803, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.03024943359196186, + "learning_rate": 8.181093625448585e-05, + "loss": 2.9165825843811035, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.030653616413474083, + "learning_rate": 8.163311885610299e-05, + "loss": 2.9171371459960938, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.03027702495455742, + "learning_rate": 8.145547980901791e-05, + "loss": 2.9399702548980713, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.030466679483652115, + "learning_rate": 8.127801925635126e-05, + "loss": 2.888066291809082, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.02997267246246338, + "learning_rate": 8.110073734107954e-05, + "loss": 2.906545877456665, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.03026283159852028, + "learning_rate": 8.092363420603584e-05, + "loss": 2.9359335899353027, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.0293166171759367, + "learning_rate": 8.074670999390896e-05, + "loss": 2.9118804931640625, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.031117571517825127, + "learning_rate": 8.056996484724342e-05, + "loss": 2.968198299407959, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.03041299246251583, + "learning_rate": 8.039339890843958e-05, + "loss": 2.9333484172821045, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.03037327155470848, + "learning_rate": 8.02170123197535e-05, + "loss": 2.895859956741333, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.03161945194005966, + "learning_rate": 8.004080522329674e-05, + "loss": 2.9491944313049316, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.031058384105563164, + "learning_rate": 7.986477776103601e-05, + "loss": 2.926038980484009, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.030037537217140198, + "learning_rate": 7.968893007479343e-05, + "loss": 2.941394805908203, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.030748063698410988, + "learning_rate": 7.951326230624658e-05, + "loss": 2.998904228210449, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.030324939638376236, + "learning_rate": 7.933777459692756e-05, + "loss": 2.9226560592651367, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.030745431780815125, + "learning_rate": 7.916246708822373e-05, + "loss": 2.977520227432251, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.02952338382601738, + "learning_rate": 7.898733992137715e-05, + "loss": 2.9365828037261963, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.030132673680782318, + "learning_rate": 7.881239323748475e-05, + "loss": 2.9574036598205566, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.030379803851246834, + "learning_rate": 7.863762717749771e-05, + "loss": 2.9329113960266113, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.029670827090740204, + "learning_rate": 7.846304188222189e-05, + "loss": 2.951117515563965, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.029823293909430504, + "learning_rate": 7.828863749231777e-05, + "loss": 2.9586987495422363, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.03546741604804993, + "learning_rate": 7.811441414829958e-05, + "loss": 2.9378662109375, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.417913436889648, + "eval_runtime": 69.6157, + "eval_samples_per_second": 35.078, + "eval_steps_per_second": 1.106, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.033859945833683014, + "learning_rate": 7.794037199053598e-05, + "loss": 2.8493661880493164, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.032931916415691376, + "learning_rate": 7.776651115924959e-05, + "loss": 2.844602108001709, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.03232948109507561, + "learning_rate": 7.759283179451704e-05, + "loss": 2.8463873863220215, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.032815027981996536, + "learning_rate": 7.741933403626848e-05, + "loss": 2.846658706665039, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.03269147500395775, + "learning_rate": 7.724601802428805e-05, + "loss": 2.8286149501800537, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.03337012976408005, + "learning_rate": 7.707288389821329e-05, + "loss": 2.9202351570129395, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.031733594834804535, + "learning_rate": 7.689993179753519e-05, + "loss": 2.8601999282836914, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.032877493649721146, + "learning_rate": 7.672716186159829e-05, + "loss": 2.8566389083862305, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.033852867782115936, + "learning_rate": 7.655457422959993e-05, + "loss": 2.859797477722168, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.03143248334527016, + "learning_rate": 7.638216904059122e-05, + "loss": 2.8411474227905273, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.033198583871126175, + "learning_rate": 7.620994643347559e-05, + "loss": 2.888702392578125, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.029946891590952873, + "learning_rate": 7.60379065470098e-05, + "loss": 2.8332901000976562, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.031660862267017365, + "learning_rate": 7.586604951980326e-05, + "loss": 2.8255481719970703, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.030189910903573036, + "learning_rate": 7.569437549031813e-05, + "loss": 2.8664190769195557, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.03273668885231018, + "learning_rate": 7.55228845968691e-05, + "loss": 2.886531114578247, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.030584124848246574, + "learning_rate": 7.535157697762313e-05, + "loss": 2.872990131378174, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.032543931156396866, + "learning_rate": 7.518045277059979e-05, + "loss": 2.891528606414795, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.031016048043966293, + "learning_rate": 7.500951211367068e-05, + "loss": 2.8759872913360596, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.03190434351563454, + "learning_rate": 7.483875514455979e-05, + "loss": 2.8757526874542236, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.030996622517704964, + "learning_rate": 7.466818200084264e-05, + "loss": 2.862654447555542, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.03289523348212242, + "learning_rate": 7.449779281994712e-05, + "loss": 2.9064857959747314, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.03322640061378479, + "learning_rate": 7.43275877391528e-05, + "loss": 2.8921048641204834, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.0303682629019022, + "learning_rate": 7.415756689559061e-05, + "loss": 2.886458396911621, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.033412594348192215, + "learning_rate": 7.398773042624341e-05, + "loss": 2.8638253211975098, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.030457086861133575, + "learning_rate": 7.381807846794532e-05, + "loss": 2.835578441619873, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.03225106745958328, + "learning_rate": 7.3648611157382e-05, + "loss": 2.8652632236480713, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.03178466483950615, + "learning_rate": 7.34793286310899e-05, + "loss": 2.8796353340148926, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.03075973503291607, + "learning_rate": 7.331023102545716e-05, + "loss": 2.895150661468506, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.02984696812927723, + "learning_rate": 7.314131847672268e-05, + "loss": 2.882878303527832, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.030945835635066032, + "learning_rate": 7.297259112097608e-05, + "loss": 2.898838520050049, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.030904723331332207, + "learning_rate": 7.280404909415801e-05, + "loss": 2.885678291320801, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.030709490180015564, + "learning_rate": 7.263569253205973e-05, + "loss": 2.8645198345184326, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.030528701841831207, + "learning_rate": 7.24675215703232e-05, + "loss": 2.8877408504486084, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.030471473932266235, + "learning_rate": 7.229953634444055e-05, + "loss": 2.9136905670166016, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.030309896916151047, + "learning_rate": 7.213173698975444e-05, + "loss": 2.9238810539245605, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.030605493113398552, + "learning_rate": 7.1964123641458e-05, + "loss": 2.8624460697174072, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.029686832800507545, + "learning_rate": 7.179669643459405e-05, + "loss": 2.9085543155670166, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.030039243400096893, + "learning_rate": 7.162945550405575e-05, + "loss": 2.8624000549316406, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.02953776717185974, + "learning_rate": 7.146240098458613e-05, + "loss": 2.8893938064575195, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.029651308432221413, + "learning_rate": 7.129553301077801e-05, + "loss": 2.914975166320801, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.030902806669473648, + "learning_rate": 7.112885171707378e-05, + "loss": 2.887071132659912, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.02948773466050625, + "learning_rate": 7.096235723776562e-05, + "loss": 2.8893041610717773, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.03049607202410698, + "learning_rate": 7.079604970699514e-05, + "loss": 2.902736186981201, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.030077926814556122, + "learning_rate": 7.062992925875332e-05, + "loss": 2.8936967849731445, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.0303986556828022, + "learning_rate": 7.046399602688045e-05, + "loss": 2.9223172664642334, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.030650658532977104, + "learning_rate": 7.029825014506572e-05, + "loss": 2.9140920639038086, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.029562659561634064, + "learning_rate": 7.013269174684795e-05, + "loss": 2.918179988861084, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.03063463792204857, + "learning_rate": 6.996732096561422e-05, + "loss": 2.9222826957702637, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.030284354463219643, + "learning_rate": 6.980213793460092e-05, + "loss": 2.8685007095336914, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.029791876673698425, + "learning_rate": 6.963714278689304e-05, + "loss": 2.914102554321289, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.030152756720781326, + "learning_rate": 6.947233565542417e-05, + "loss": 2.8890669345855713, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.02957645244896412, + "learning_rate": 6.930771667297653e-05, + "loss": 2.8851962089538574, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.02963421493768692, + "learning_rate": 6.914328597218042e-05, + "loss": 2.9012060165405273, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.029279837384819984, + "learning_rate": 6.8979043685515e-05, + "loss": 2.949984073638916, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.0302849430590868, + "learning_rate": 6.881498994530708e-05, + "loss": 2.9049415588378906, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.029432371258735657, + "learning_rate": 6.865112488373186e-05, + "loss": 2.8919944763183594, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.030542058870196342, + "learning_rate": 6.84874486328125e-05, + "loss": 2.9168801307678223, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.02911970019340515, + "learning_rate": 6.832396132441993e-05, + "loss": 2.9342129230499268, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.030335791409015656, + "learning_rate": 6.816066309027309e-05, + "loss": 2.918611764907837, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.03031272068619728, + "learning_rate": 6.799755406193815e-05, + "loss": 2.8769750595092773, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.030200788751244545, + "learning_rate": 6.783463437082929e-05, + "loss": 2.8994460105895996, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.03032573312520981, + "learning_rate": 6.767190414820792e-05, + "loss": 2.920720100402832, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.03180326521396637, + "learning_rate": 6.750936352518284e-05, + "loss": 2.891145706176758, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.029451677575707436, + "learning_rate": 6.734701263271011e-05, + "loss": 2.9057083129882812, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.030261319130659103, + "learning_rate": 6.718485160159289e-05, + "loss": 2.9132461547851562, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.031257566064596176, + "learning_rate": 6.702288056248145e-05, + "loss": 2.9089136123657227, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.03043205663561821, + "learning_rate": 6.686109964587285e-05, + "loss": 2.912537097930908, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.030708536505699158, + "learning_rate": 6.669950898211106e-05, + "loss": 2.8841137886047363, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.03044111281633377, + "learning_rate": 6.653810870138676e-05, + "loss": 2.921257257461548, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.029576802626252174, + "learning_rate": 6.637689893373729e-05, + "loss": 2.908468723297119, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.03031427040696144, + "learning_rate": 6.621587980904651e-05, + "loss": 2.936445713043213, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.036312974989414215, + "learning_rate": 6.605505145704437e-05, + "loss": 2.8997855186462402, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.443755149841309, + "eval_runtime": 62.9898, + "eval_samples_per_second": 38.768, + "eval_steps_per_second": 1.222, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.03337511047720909, + "learning_rate": 6.58944140073077e-05, + "loss": 2.8584513664245605, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.031415343284606934, + "learning_rate": 6.573396758925891e-05, + "loss": 2.8529202938079834, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.033063437789678574, + "learning_rate": 6.557371233216692e-05, + "loss": 2.814058780670166, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.03153088316321373, + "learning_rate": 6.541364836514646e-05, + "loss": 2.80460786819458, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.03227711841464043, + "learning_rate": 6.525377581715829e-05, + "loss": 2.858191967010498, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.030423127114772797, + "learning_rate": 6.509409481700864e-05, + "loss": 2.845815420150757, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.032030411064624786, + "learning_rate": 6.49346054933496e-05, + "loss": 2.837918281555176, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.030464086681604385, + "learning_rate": 6.477530797467911e-05, + "loss": 2.841099739074707, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.03194798156619072, + "learning_rate": 6.461620238934006e-05, + "loss": 2.843568801879883, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.03152863308787346, + "learning_rate": 6.445728886552109e-05, + "loss": 2.8779497146606445, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.03201183304190636, + "learning_rate": 6.429856753125573e-05, + "loss": 2.906585216522217, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.02999979443848133, + "learning_rate": 6.414003851442318e-05, + "loss": 2.84859561920166, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.03106156550347805, + "learning_rate": 6.398170194274722e-05, + "loss": 2.814643383026123, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.03118651732802391, + "learning_rate": 6.38235579437968e-05, + "loss": 2.8370001316070557, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.031022636219859123, + "learning_rate": 6.366560664498572e-05, + "loss": 2.8466384410858154, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.03163835406303406, + "learning_rate": 6.350784817357242e-05, + "loss": 2.808440685272217, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.03071477636694908, + "learning_rate": 6.335028265666023e-05, + "loss": 2.850621223449707, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.03041849471628666, + "learning_rate": 6.319291022119653e-05, + "loss": 2.818002939224243, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.03104640357196331, + "learning_rate": 6.303573099397378e-05, + "loss": 2.8375182151794434, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.02970169298350811, + "learning_rate": 6.287874510162821e-05, + "loss": 2.871188163757324, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.03146181255578995, + "learning_rate": 6.272195267064063e-05, + "loss": 2.8356661796569824, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.030204759910702705, + "learning_rate": 6.25653538273358e-05, + "loss": 2.861675262451172, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.029957789927721024, + "learning_rate": 6.240894869788267e-05, + "loss": 2.8493146896362305, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.031224241480231285, + "learning_rate": 6.225273740829404e-05, + "loss": 2.876760959625244, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 0.02914278209209442, + "learning_rate": 6.209672008442635e-05, + "loss": 2.838275909423828, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.03204638510942459, + "learning_rate": 6.19408968519801e-05, + "loss": 2.8679757118225098, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.030257491394877434, + "learning_rate": 6.178526783649916e-05, + "loss": 2.8245105743408203, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.031307291239500046, + "learning_rate": 6.162983316337109e-05, + "loss": 2.852024793624878, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 0.030594198033213615, + "learning_rate": 6.147459295782676e-05, + "loss": 2.883610248565674, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.03119277022778988, + "learning_rate": 6.131954734494045e-05, + "loss": 2.866126775741577, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.0302574560046196, + "learning_rate": 6.11646964496296e-05, + "loss": 2.857886791229248, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.029702702537178993, + "learning_rate": 6.101004039665471e-05, + "loss": 2.8324077129364014, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.029572894796729088, + "learning_rate": 6.085557931061937e-05, + "loss": 2.878951072692871, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.03024783544242382, + "learning_rate": 6.070131331597015e-05, + "loss": 2.900284767150879, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.02952091582119465, + "learning_rate": 6.054724253699636e-05, + "loss": 2.8402292728424072, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.03006756864488125, + "learning_rate": 6.03933670978301e-05, + "loss": 2.8609299659729004, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.029146816581487656, + "learning_rate": 6.0239687122445826e-05, + "loss": 2.887960433959961, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.03026733547449112, + "learning_rate": 6.0086202734661e-05, + "loss": 2.852726459503174, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.029839467257261276, + "learning_rate": 5.9932914058135004e-05, + "loss": 2.8559489250183105, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.029919864609837532, + "learning_rate": 5.977982121636984e-05, + "loss": 2.880255937576294, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.03074759431183338, + "learning_rate": 5.962692433270962e-05, + "loss": 2.8586673736572266, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.029141079634428024, + "learning_rate": 5.9474223530340634e-05, + "loss": 2.8813042640686035, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.031023059040308, + "learning_rate": 5.932171893229124e-05, + "loss": 2.8355565071105957, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.029786985367536545, + "learning_rate": 5.916941066143137e-05, + "loss": 2.872175693511963, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.030190780758857727, + "learning_rate": 5.9017298840473364e-05, + "loss": 2.874236583709717, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.029669199138879776, + "learning_rate": 5.8865383591970775e-05, + "loss": 2.8723950386047363, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.030073825269937515, + "learning_rate": 5.8713665038319e-05, + "loss": 2.863720417022705, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.02976175583899021, + "learning_rate": 5.856214330175498e-05, + "loss": 2.883267402648926, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.030071711167693138, + "learning_rate": 5.841081850435704e-05, + "loss": 2.8690664768218994, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.031068995594978333, + "learning_rate": 5.825969076804488e-05, + "loss": 2.9107182025909424, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.030393972992897034, + "learning_rate": 5.810876021457925e-05, + "loss": 2.859959602355957, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.030435480177402496, + "learning_rate": 5.795802696556222e-05, + "loss": 2.9003896713256836, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.02982497587800026, + "learning_rate": 5.78074911424369e-05, + "loss": 2.857231378555298, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.030999863520264626, + "learning_rate": 5.765715286648734e-05, + "loss": 2.8567168712615967, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.029698362573981285, + "learning_rate": 5.7507012258838154e-05, + "loss": 2.8949484825134277, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.031584806740283966, + "learning_rate": 5.7357069440455095e-05, + "loss": 2.8703250885009766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.02952221967279911, + "learning_rate": 5.7207324532144454e-05, + "loss": 2.8914244174957275, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.03138599172234535, + "learning_rate": 5.7057777654552785e-05, + "loss": 2.917753219604492, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.029678434133529663, + "learning_rate": 5.690842892816741e-05, + "loss": 2.8854236602783203, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.03174183890223503, + "learning_rate": 5.675927847331593e-05, + "loss": 2.8508100509643555, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.03006329946219921, + "learning_rate": 5.661032641016619e-05, + "loss": 2.883819103240967, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.031188689172267914, + "learning_rate": 5.6461572858726086e-05, + "loss": 2.932220458984375, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.029729023575782776, + "learning_rate": 5.6313017938843595e-05, + "loss": 2.894927978515625, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.02953082136809826, + "learning_rate": 5.6164661770207026e-05, + "loss": 2.896371364593506, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.030423447489738464, + "learning_rate": 5.6016504472344e-05, + "loss": 2.9003541469573975, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.029901325702667236, + "learning_rate": 5.5868546164622284e-05, + "loss": 2.8726208209991455, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.030071355402469635, + "learning_rate": 5.572078696624917e-05, + "loss": 2.881291151046753, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.029847780242562294, + "learning_rate": 5.557322699627168e-05, + "loss": 2.944478988647461, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.029794953763484955, + "learning_rate": 5.542586637357607e-05, + "loss": 2.8625364303588867, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.03029358573257923, + "learning_rate": 5.527870521688817e-05, + "loss": 2.9106764793395996, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.031095417216420174, + "learning_rate": 5.513174364477311e-05, + "loss": 2.8879761695861816, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.03671329468488693, + "learning_rate": 5.498498177563518e-05, + "loss": 2.877009391784668, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 4.470407962799072, + "eval_runtime": 57.8161, + "eval_samples_per_second": 42.237, + "eval_steps_per_second": 1.332, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.03383656591176987, + "learning_rate": 5.4838419727717716e-05, + "loss": 2.8301892280578613, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.030587317422032356, + "learning_rate": 5.4692057619103144e-05, + "loss": 2.809070587158203, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.034895412623882294, + "learning_rate": 5.454589556771286e-05, + "loss": 2.845365285873413, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.03129304572939873, + "learning_rate": 5.439993369130681e-05, + "loss": 2.8176681995391846, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.03369338437914848, + "learning_rate": 5.4254172107483924e-05, + "loss": 2.820741653442383, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.030895931646227837, + "learning_rate": 5.410861093368169e-05, + "loss": 2.810272693634033, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.03239824250340462, + "learning_rate": 5.3963250287176094e-05, + "loss": 2.855811595916748, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.03123638592660427, + "learning_rate": 5.381809028508168e-05, + "loss": 2.7941160202026367, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.031117001548409462, + "learning_rate": 5.367313104435104e-05, + "loss": 2.784010887145996, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.03162471950054169, + "learning_rate": 5.352837268177544e-05, + "loss": 2.8337666988372803, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.0316263884305954, + "learning_rate": 5.3383815313983916e-05, + "loss": 2.8019790649414062, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.031872715801000595, + "learning_rate": 5.323945905744379e-05, + "loss": 2.8156003952026367, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.030252376571297646, + "learning_rate": 5.3095304028460284e-05, + "loss": 2.7971153259277344, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.03168163076043129, + "learning_rate": 5.2951350343176495e-05, + "loss": 2.791532039642334, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.03014831617474556, + "learning_rate": 5.2807598117573384e-05, + "loss": 2.8359432220458984, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.03142879158258438, + "learning_rate": 5.266404746746939e-05, + "loss": 2.846064567565918, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.030883461236953735, + "learning_rate": 5.252069850852072e-05, + "loss": 2.8668103218078613, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.03189866617321968, + "learning_rate": 5.237755135622107e-05, + "loss": 2.8214638233184814, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.029116913676261902, + "learning_rate": 5.223460612590159e-05, + "loss": 2.850940704345703, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.030518673360347748, + "learning_rate": 5.209186293273044e-05, + "loss": 2.844576358795166, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.029227688908576965, + "learning_rate": 5.1949321891713446e-05, + "loss": 2.797168493270874, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.029839880764484406, + "learning_rate": 5.180698311769338e-05, + "loss": 2.8534903526306152, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.03011847473680973, + "learning_rate": 5.166484672534983e-05, + "loss": 2.8278157711029053, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.03027244098484516, + "learning_rate": 5.152291282919961e-05, + "loss": 2.858158588409424, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.030171507969498634, + "learning_rate": 5.138118154359632e-05, + "loss": 2.8523292541503906, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.029657812789082527, + "learning_rate": 5.1239652982730335e-05, + "loss": 2.8015360832214355, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.028928041458129883, + "learning_rate": 5.1098327260628466e-05, + "loss": 2.8357324600219727, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.029674772173166275, + "learning_rate": 5.09572044911545e-05, + "loss": 2.8269171714782715, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.03013371117413044, + "learning_rate": 5.0816284788008527e-05, + "loss": 2.822054862976074, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.02960435301065445, + "learning_rate": 5.067556826472684e-05, + "loss": 2.8452866077423096, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.030388740822672844, + "learning_rate": 5.053505503468228e-05, + "loss": 2.855799674987793, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.029823970049619675, + "learning_rate": 5.0394745211083785e-05, + "loss": 2.8262312412261963, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.030613111332058907, + "learning_rate": 5.025463890697655e-05, + "loss": 2.8490824699401855, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.029467277228832245, + "learning_rate": 5.011473623524159e-05, + "loss": 2.855149984359741, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.030264565721154213, + "learning_rate": 4.9975037308595864e-05, + "loss": 2.849930763244629, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.02961486577987671, + "learning_rate": 4.983554223959257e-05, + "loss": 2.843088150024414, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.029051944613456726, + "learning_rate": 4.9696251140620134e-05, + "loss": 2.825998067855835, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.03038879856467247, + "learning_rate": 4.9557164123902924e-05, + "loss": 2.8565316200256348, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.02917695790529251, + "learning_rate": 4.9418281301500844e-05, + "loss": 2.8733534812927246, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.03011898323893547, + "learning_rate": 4.9279602785309365e-05, + "loss": 2.8347158432006836, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.03015534020960331, + "learning_rate": 4.914112868705911e-05, + "loss": 2.878455400466919, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.02992299199104309, + "learning_rate": 4.900285911831624e-05, + "loss": 2.861504554748535, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.030217986553907394, + "learning_rate": 4.886479419048201e-05, + "loss": 2.8451390266418457, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.029564134776592255, + "learning_rate": 4.872693401479292e-05, + "loss": 2.8065595626831055, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.03130771592259407, + "learning_rate": 4.858927870232039e-05, + "loss": 2.8506879806518555, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.030411556363105774, + "learning_rate": 4.845182836397068e-05, + "loss": 2.857304573059082, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.030868662521243095, + "learning_rate": 4.831458311048532e-05, + "loss": 2.8545470237731934, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.03002075105905533, + "learning_rate": 4.817754305244008e-05, + "loss": 2.889376640319824, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.030850842595100403, + "learning_rate": 4.804070830024577e-05, + "loss": 2.8634769916534424, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.03078618459403515, + "learning_rate": 4.7904078964147654e-05, + "loss": 2.8487653732299805, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.031029092147946358, + "learning_rate": 4.776765515422557e-05, + "loss": 2.897327184677124, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.03101371042430401, + "learning_rate": 4.763143698039371e-05, + "loss": 2.8642020225524902, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.03044370748102665, + "learning_rate": 4.749542455240043e-05, + "loss": 2.87313175201416, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.030529852956533432, + "learning_rate": 4.735961797982876e-05, + "loss": 2.849147081375122, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.029675021767616272, + "learning_rate": 4.7224017372095414e-05, + "loss": 2.8475494384765625, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.03002132847905159, + "learning_rate": 4.708862283845143e-05, + "loss": 2.8713879585266113, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.03048008307814598, + "learning_rate": 4.695343448798163e-05, + "loss": 2.8976492881774902, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.02930760383605957, + "learning_rate": 4.681845242960492e-05, + "loss": 2.850775718688965, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.029840776696801186, + "learning_rate": 4.668367677207398e-05, + "loss": 2.8743128776550293, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.030162667855620384, + "learning_rate": 4.654910762397499e-05, + "loss": 2.850478410720825, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.029736310243606567, + "learning_rate": 4.6414745093727913e-05, + "loss": 2.851759195327759, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.029385298490524292, + "learning_rate": 4.6280589289586255e-05, + "loss": 2.8473241329193115, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.028572317212820053, + "learning_rate": 4.614664031963692e-05, + "loss": 2.852155923843384, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.0300681721419096, + "learning_rate": 4.601289829180004e-05, + "loss": 2.871021270751953, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.0292969960719347, + "learning_rate": 4.587936331382934e-05, + "loss": 2.8470516204833984, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.029493169859051704, + "learning_rate": 4.574603549331151e-05, + "loss": 2.892655372619629, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.029351307079195976, + "learning_rate": 4.561291493766625e-05, + "loss": 2.8886585235595703, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.029827935621142387, + "learning_rate": 4.5480001754146455e-05, + "loss": 2.8802051544189453, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.029231170192360878, + "learning_rate": 4.5347296049837875e-05, + "loss": 2.876192092895508, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.029604196548461914, + "learning_rate": 4.52147979316592e-05, + "loss": 2.8420169353485107, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.030229991301894188, + "learning_rate": 4.5082507506361574e-05, + "loss": 2.86871075630188, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.0357736200094223, + "learning_rate": 4.495042488052904e-05, + "loss": 2.852806568145752, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 4.496274948120117, + "eval_runtime": 58.0976, + "eval_samples_per_second": 42.033, + "eval_steps_per_second": 1.325, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.03180048242211342, + "learning_rate": 4.481855016057839e-05, + "loss": 2.85683012008667, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.030023347586393356, + "learning_rate": 4.468688345275848e-05, + "loss": 2.814507484436035, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.03346574679017067, + "learning_rate": 4.455542486315086e-05, + "loss": 2.8209877014160156, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.030521582812070847, + "learning_rate": 4.442417449766938e-05, + "loss": 2.7859511375427246, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.033295996487140656, + "learning_rate": 4.429313246206014e-05, + "loss": 2.7879018783569336, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.029781939461827278, + "learning_rate": 4.416229886190119e-05, + "loss": 2.801133155822754, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.033207762986421585, + "learning_rate": 4.403167380260281e-05, + "loss": 2.8046321868896484, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.03027796931564808, + "learning_rate": 4.3901257389407475e-05, + "loss": 2.8235392570495605, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.03096267767250538, + "learning_rate": 4.3771049727389075e-05, + "loss": 2.8181068897247314, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.030886022374033928, + "learning_rate": 4.364105092145377e-05, + "loss": 2.7954399585723877, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.030112111940979958, + "learning_rate": 4.3511261076339e-05, + "loss": 2.78225040435791, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.03067166730761528, + "learning_rate": 4.3381680296614394e-05, + "loss": 2.8147826194763184, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.029821906238794327, + "learning_rate": 4.3252308686680626e-05, + "loss": 2.811027765274048, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.030433299019932747, + "learning_rate": 4.3123146350770146e-05, + "loss": 2.8178303241729736, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.02960469201207161, + "learning_rate": 4.2994193392946726e-05, + "loss": 2.802370548248291, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.029496584087610245, + "learning_rate": 4.286544991710541e-05, + "loss": 2.8156142234802246, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.029699044302105904, + "learning_rate": 4.2736916026972576e-05, + "loss": 2.7911806106567383, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.030225902795791626, + "learning_rate": 4.260859182610542e-05, + "loss": 2.7872819900512695, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.02939198724925518, + "learning_rate": 4.2480477417892776e-05, + "loss": 2.800001621246338, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.03059281036257744, + "learning_rate": 4.235257290555381e-05, + "loss": 2.8127222061157227, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.02927992306649685, + "learning_rate": 4.222487839213903e-05, + "loss": 2.8320517539978027, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.03038649819791317, + "learning_rate": 4.209739398052956e-05, + "loss": 2.8169116973876953, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.029806654900312424, + "learning_rate": 4.1970119773437316e-05, + "loss": 2.8279008865356445, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.029620900750160217, + "learning_rate": 4.184305587340483e-05, + "loss": 2.8212599754333496, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.029498178511857986, + "learning_rate": 4.171620238280511e-05, + "loss": 2.8112261295318604, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.02918926067650318, + "learning_rate": 4.158955940384179e-05, + "loss": 2.8290767669677734, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.029827842488884926, + "learning_rate": 4.146312703854881e-05, + "loss": 2.8396403789520264, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.029516329988837242, + "learning_rate": 4.133690538879046e-05, + "loss": 2.840583324432373, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.029057780280709267, + "learning_rate": 4.1210894556261226e-05, + "loss": 2.835113525390625, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.029345065355300903, + "learning_rate": 4.108509464248578e-05, + "loss": 2.8197970390319824, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.02873479388654232, + "learning_rate": 4.095950574881891e-05, + "loss": 2.8052587509155273, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.029737083241343498, + "learning_rate": 4.083412797644514e-05, + "loss": 2.819146156311035, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.0292612686753273, + "learning_rate": 4.0708961426379214e-05, + "loss": 2.8543701171875, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.028945421800017357, + "learning_rate": 4.0584006199465484e-05, + "loss": 2.804084300994873, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.029137294739484787, + "learning_rate": 4.0459262396378165e-05, + "loss": 2.7658708095550537, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.02911440096795559, + "learning_rate": 4.033473011762116e-05, + "loss": 2.8134429454803467, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.028932519257068634, + "learning_rate": 4.0210409463527656e-05, + "loss": 2.810455322265625, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.029404157772660255, + "learning_rate": 4.008630053426082e-05, + "loss": 2.802886962890625, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.028970446437597275, + "learning_rate": 3.996240342981279e-05, + "loss": 2.8426849842071533, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.029187830165028572, + "learning_rate": 3.983871825000529e-05, + "loss": 2.8366644382476807, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.029293285682797432, + "learning_rate": 3.971524509448925e-05, + "loss": 2.8261566162109375, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.029366638511419296, + "learning_rate": 3.959198406274472e-05, + "loss": 2.8125827312469482, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.02900587022304535, + "learning_rate": 3.946893525408097e-05, + "loss": 2.839684009552002, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.029162462800741196, + "learning_rate": 3.934609876763604e-05, + "loss": 2.8331997394561768, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.029060618951916695, + "learning_rate": 3.922347470237728e-05, + "loss": 2.8627192974090576, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.02887798845767975, + "learning_rate": 3.910106315710053e-05, + "loss": 2.859687089920044, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.02894580364227295, + "learning_rate": 3.897886423043061e-05, + "loss": 2.8195912837982178, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.029554955661296844, + "learning_rate": 3.8856878020820975e-05, + "loss": 2.852871894836426, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.02822914719581604, + "learning_rate": 3.873510462655375e-05, + "loss": 2.8671388626098633, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.02954651229083538, + "learning_rate": 3.861354414573954e-05, + "loss": 2.8524365425109863, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.02902315929532051, + "learning_rate": 3.849219667631735e-05, + "loss": 2.843954563140869, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.0290356632322073, + "learning_rate": 3.8371062316054764e-05, + "loss": 2.8504233360290527, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.029262971132993698, + "learning_rate": 3.8250141162547466e-05, + "loss": 2.8657913208007812, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.029469065368175507, + "learning_rate": 3.812943331321956e-05, + "loss": 2.82297945022583, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.029609832912683487, + "learning_rate": 3.8008938865322955e-05, + "loss": 2.856297016143799, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.029690174385905266, + "learning_rate": 3.788865791593813e-05, + "loss": 2.8302533626556396, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.03018389455974102, + "learning_rate": 3.7768590561973064e-05, + "loss": 2.8542540073394775, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.029053693637251854, + "learning_rate": 3.764873690016394e-05, + "loss": 2.8432576656341553, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.028549278154969215, + "learning_rate": 3.7529097027074654e-05, + "loss": 2.8324697017669678, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.030014613643288612, + "learning_rate": 3.7409671039096936e-05, + "loss": 2.800536870956421, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.02949996292591095, + "learning_rate": 3.7290459032450184e-05, + "loss": 2.835141897201538, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.028825247660279274, + "learning_rate": 3.717146110318123e-05, + "loss": 2.8263447284698486, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.029436098411679268, + "learning_rate": 3.7052677347164614e-05, + "loss": 2.8522257804870605, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.029510991647839546, + "learning_rate": 3.693410786010231e-05, + "loss": 2.851526975631714, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.028975380584597588, + "learning_rate": 3.6815752737523537e-05, + "loss": 2.820462465286255, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.029128702357411385, + "learning_rate": 3.6697612074784944e-05, + "loss": 2.861868381500244, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.02951551228761673, + "learning_rate": 3.657968596707029e-05, + "loss": 2.875210762023926, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.028196118772029877, + "learning_rate": 3.64619745093906e-05, + "loss": 2.8345048427581787, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.029215430840849876, + "learning_rate": 3.6344477796583714e-05, + "loss": 2.84328556060791, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.02857460267841816, + "learning_rate": 3.622719592331472e-05, + "loss": 2.8448305130004883, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.02842608466744423, + "learning_rate": 3.611012898407544e-05, + "loss": 2.860792636871338, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.03630338981747627, + "learning_rate": 3.599327707318464e-05, + "loss": 2.8723089694976807, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 4.522201061248779, + "eval_runtime": 57.3764, + "eval_samples_per_second": 42.561, + "eval_steps_per_second": 1.342, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.030005959793925285, + "learning_rate": 3.587664028478779e-05, + "loss": 2.7629683017730713, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.028917059302330017, + "learning_rate": 3.5760218712857046e-05, + "loss": 2.7894129753112793, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.029286719858646393, + "learning_rate": 3.564401245119119e-05, + "loss": 2.8119699954986572, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.02967856638133526, + "learning_rate": 3.552802159341543e-05, + "loss": 2.8115768432617188, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.029460743069648743, + "learning_rate": 3.541224623298155e-05, + "loss": 2.7809553146362305, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.028186975046992302, + "learning_rate": 3.529668646316765e-05, + "loss": 2.818805456161499, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.02975514344871044, + "learning_rate": 3.5181342377078166e-05, + "loss": 2.816215991973877, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.029532121494412422, + "learning_rate": 3.5066214067643785e-05, + "loss": 2.782461166381836, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.02879161946475506, + "learning_rate": 3.495130162762119e-05, + "loss": 2.768501043319702, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.029776159673929214, + "learning_rate": 3.483660514959343e-05, + "loss": 2.7995424270629883, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.028667304664850235, + "learning_rate": 3.4722124725969235e-05, + "loss": 2.773409843444824, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.028793245553970337, + "learning_rate": 3.460786044898346e-05, + "loss": 2.7938318252563477, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.029090600088238716, + "learning_rate": 3.4493812410696784e-05, + "loss": 2.823000907897949, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.028883792459964752, + "learning_rate": 3.437998070299564e-05, + "loss": 2.8175344467163086, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.028518877923488617, + "learning_rate": 3.426636541759225e-05, + "loss": 2.792815685272217, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.029769321903586388, + "learning_rate": 3.415296664602425e-05, + "loss": 2.831251382827759, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.029009172692894936, + "learning_rate": 3.403978447965507e-05, + "loss": 2.8345084190368652, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.028994156047701836, + "learning_rate": 3.392681900967355e-05, + "loss": 2.7925689220428467, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.029505716636776924, + "learning_rate": 3.381407032709395e-05, + "loss": 2.814758777618408, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.028683487325906754, + "learning_rate": 3.3701538522755714e-05, + "loss": 2.7954044342041016, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.029633600264787674, + "learning_rate": 3.35892236873238e-05, + "loss": 2.794569969177246, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.028943300247192383, + "learning_rate": 3.3477125911288284e-05, + "loss": 2.803800106048584, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.029026521369814873, + "learning_rate": 3.3365245284964225e-05, + "loss": 2.760383129119873, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.029057513922452927, + "learning_rate": 3.3253581898491836e-05, + "loss": 2.8081226348876953, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.02906658686697483, + "learning_rate": 3.3142135841836276e-05, + "loss": 2.8183770179748535, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.02926407754421234, + "learning_rate": 3.303090720478773e-05, + "loss": 2.831427574157715, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.029171453788876534, + "learning_rate": 3.291989607696092e-05, + "loss": 2.826587200164795, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.029656723141670227, + "learning_rate": 3.280910254779565e-05, + "loss": 2.7633590698242188, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.029490994289517403, + "learning_rate": 3.269852670655629e-05, + "loss": 2.7932424545288086, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.0291811041533947, + "learning_rate": 3.258816864233169e-05, + "loss": 2.8346633911132812, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.029164446517825127, + "learning_rate": 3.247802844403536e-05, + "loss": 2.8193140029907227, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.029394259676337242, + "learning_rate": 3.236810620040532e-05, + "loss": 2.8272342681884766, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.029221080243587494, + "learning_rate": 3.225840200000396e-05, + "loss": 2.79146146774292, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.028165455907583237, + "learning_rate": 3.214891593121788e-05, + "loss": 2.7826766967773438, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.0296048391610384, + "learning_rate": 3.203964808225801e-05, + "loss": 2.8119699954986572, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.028316112235188484, + "learning_rate": 3.193059854115968e-05, + "loss": 2.824860095977783, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02878737449645996, + "learning_rate": 3.182176739578195e-05, + "loss": 2.811997890472412, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029061628505587578, + "learning_rate": 3.171315473380815e-05, + "loss": 2.8370048999786377, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028607595711946487, + "learning_rate": 3.160476064274555e-05, + "loss": 2.8093791007995605, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.028144560754299164, + "learning_rate": 3.149658520992537e-05, + "loss": 2.8236052989959717, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.028813332319259644, + "learning_rate": 3.138862852250249e-05, + "loss": 2.838690996170044, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.028437303379178047, + "learning_rate": 3.1280890667455694e-05, + "loss": 2.800854206085205, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.028567243367433548, + "learning_rate": 3.1173371731587496e-05, + "loss": 2.8301069736480713, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.029187796637415886, + "learning_rate": 3.1066071801523905e-05, + "loss": 2.823208808898926, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.028203707188367844, + "learning_rate": 3.095899096371456e-05, + "loss": 2.804643154144287, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.02822236344218254, + "learning_rate": 3.0852129304432506e-05, + "loss": 2.8132784366607666, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.028921568766236305, + "learning_rate": 3.074548690977434e-05, + "loss": 2.8215417861938477, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.02867986261844635, + "learning_rate": 3.0639063865659834e-05, + "loss": 2.8318984508514404, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.028849760070443153, + "learning_rate": 3.0532860257832144e-05, + "loss": 2.8209228515625, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.028631621971726418, + "learning_rate": 3.04268761718576e-05, + "loss": 2.8219287395477295, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.028188595548272133, + "learning_rate": 3.0321111693125648e-05, + "loss": 2.814027786254883, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.028504883870482445, + "learning_rate": 3.0215566906848906e-05, + "loss": 2.847404956817627, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.02826615609228611, + "learning_rate": 3.0110241898062726e-05, + "loss": 2.808562755584717, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02844209223985672, + "learning_rate": 3.0005136751625748e-05, + "loss": 2.827244520187378, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.028765909373760223, + "learning_rate": 2.9900251552219178e-05, + "loss": 2.81117844581604, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.028130069375038147, + "learning_rate": 2.9795586384347202e-05, + "loss": 2.852351188659668, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.028408419340848923, + "learning_rate": 2.9691141332336494e-05, + "loss": 2.847090244293213, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.028523078188300133, + "learning_rate": 2.9586916480336735e-05, + "loss": 2.821716785430908, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.028461646288633347, + "learning_rate": 2.9482911912320006e-05, + "loss": 2.816019296646118, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.028531840071082115, + "learning_rate": 2.9379127712080756e-05, + "loss": 2.8226165771484375, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.028871113434433937, + "learning_rate": 2.9275563963236107e-05, + "loss": 2.831231117248535, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.0285959430038929, + "learning_rate": 2.9172220749225548e-05, + "loss": 2.8257393836975098, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02841980569064617, + "learning_rate": 2.906909815331085e-05, + "loss": 2.8298444747924805, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.028909973800182343, + "learning_rate": 2.896619625857588e-05, + "loss": 2.808438777923584, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.02884981967508793, + "learning_rate": 2.8863515147927003e-05, + "loss": 2.815157890319824, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.02836848609149456, + "learning_rate": 2.8761054904092514e-05, + "loss": 2.792750835418701, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.028464308008551598, + "learning_rate": 2.8658815609622665e-05, + "loss": 2.7594008445739746, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.02909877710044384, + "learning_rate": 2.855679734688993e-05, + "loss": 2.844888925552368, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.0285524632781744, + "learning_rate": 2.8455000198088502e-05, + "loss": 2.8162052631378174, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.02916364185512066, + "learning_rate": 2.8353424245234582e-05, + "loss": 2.8175275325775146, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.028360974043607712, + "learning_rate": 2.825206957016599e-05, + "loss": 2.8322806358337402, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.0366419292986393, + "learning_rate": 2.8150936254542376e-05, + "loss": 2.8780605792999268, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 4.539337635040283, + "eval_runtime": 57.482, + "eval_samples_per_second": 42.483, + "eval_steps_per_second": 1.34, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030007537454366684, + "learning_rate": 2.8050024379845118e-05, + "loss": 2.7774059772491455, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.028395716100931168, + "learning_rate": 2.7949334027376986e-05, + "loss": 2.8240408897399902, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.029259607195854187, + "learning_rate": 2.7848865278262427e-05, + "loss": 2.7780704498291016, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.029434163123369217, + "learning_rate": 2.7748618213447268e-05, + "loss": 2.802172899246216, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.02914051152765751, + "learning_rate": 2.764859291369885e-05, + "loss": 2.746731758117676, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.02971787191927433, + "learning_rate": 2.7548789459605603e-05, + "loss": 2.7998170852661133, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.029050175100564957, + "learning_rate": 2.744920793157743e-05, + "loss": 2.786863327026367, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.030421903356909752, + "learning_rate": 2.7349848409845462e-05, + "loss": 2.789545774459839, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.029742848128080368, + "learning_rate": 2.725071097446174e-05, + "loss": 2.827561855316162, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.03001682087779045, + "learning_rate": 2.715179570529959e-05, + "loss": 2.8226547241210938, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.030296841636300087, + "learning_rate": 2.7053102682053134e-05, + "loss": 2.795665740966797, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.02908152900636196, + "learning_rate": 2.6954631984237766e-05, + "loss": 2.7598729133605957, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.029059268534183502, + "learning_rate": 2.6856383691189347e-05, + "loss": 2.786234140396118, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.02878507599234581, + "learning_rate": 2.675835788206485e-05, + "loss": 2.7754065990448, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.02881552092730999, + "learning_rate": 2.6660554635841912e-05, + "loss": 2.779616355895996, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.02847588248550892, + "learning_rate": 2.6562974031318788e-05, + "loss": 2.7556540966033936, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.028446536511182785, + "learning_rate": 2.6465616147114485e-05, + "loss": 2.7858777046203613, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.028488771989941597, + "learning_rate": 2.6368481061668344e-05, + "loss": 2.8308897018432617, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.028518928214907646, + "learning_rate": 2.6271568853240557e-05, + "loss": 2.798438310623169, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.02807544358074665, + "learning_rate": 2.6174879599911347e-05, + "loss": 2.7854204177856445, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.028449799865484238, + "learning_rate": 2.6078413379581627e-05, + "loss": 2.787806510925293, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.02806367352604866, + "learning_rate": 2.5982170269972397e-05, + "loss": 2.773944139480591, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.028127942234277725, + "learning_rate": 2.588615034862507e-05, + "loss": 2.795468807220459, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.028206733986735344, + "learning_rate": 2.5790353692901158e-05, + "loss": 2.794552803039551, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.02769204042851925, + "learning_rate": 2.5694780379982238e-05, + "loss": 2.78456974029541, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.028235256671905518, + "learning_rate": 2.5599430486870055e-05, + "loss": 2.8258988857269287, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.028410933911800385, + "learning_rate": 2.5504304090386225e-05, + "loss": 2.8145394325256348, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.028073569759726524, + "learning_rate": 2.5409401267172474e-05, + "loss": 2.739804744720459, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.027985481545329094, + "learning_rate": 2.5314722093690176e-05, + "loss": 2.7385730743408203, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.0285568255931139, + "learning_rate": 2.522026664622071e-05, + "loss": 2.790156364440918, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.028381718322634697, + "learning_rate": 2.5126035000865124e-05, + "loss": 2.7715389728546143, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028455479070544243, + "learning_rate": 2.503202723354413e-05, + "loss": 2.804957628250122, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.028330320492386818, + "learning_rate": 2.4938243419998047e-05, + "loss": 2.813432216644287, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.02832239493727684, + "learning_rate": 2.484468363578689e-05, + "loss": 2.8087422847747803, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.028423218056559563, + "learning_rate": 2.475134795629004e-05, + "loss": 2.790351390838623, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.028114303946495056, + "learning_rate": 2.4658236456706357e-05, + "loss": 2.8367419242858887, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.028525900095701218, + "learning_rate": 2.456534921205403e-05, + "loss": 2.8024988174438477, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.027810173109173775, + "learning_rate": 2.4472686297170822e-05, + "loss": 2.8156440258026123, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.028172023594379425, + "learning_rate": 2.4380247786713403e-05, + "loss": 2.801828384399414, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.027999311685562134, + "learning_rate": 2.428803375515788e-05, + "loss": 2.7917897701263428, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.02863648533821106, + "learning_rate": 2.4196044276799397e-05, + "loss": 2.7998974323272705, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.027806704863905907, + "learning_rate": 2.410427942575228e-05, + "loss": 2.7955031394958496, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.028083806857466698, + "learning_rate": 2.4012739275949746e-05, + "loss": 2.8160462379455566, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02817711979150772, + "learning_rate": 2.3921423901144008e-05, + "loss": 2.7673726081848145, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.027995774522423744, + "learning_rate": 2.3830333374906342e-05, + "loss": 2.787623882293701, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.028293436393141747, + "learning_rate": 2.3739467770626637e-05, + "loss": 2.848489284515381, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.028710173442959785, + "learning_rate": 2.3648827161513686e-05, + "loss": 2.8186140060424805, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.028271062299609184, + "learning_rate": 2.3558411620595038e-05, + "loss": 2.822369337081909, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.02830388769507408, + "learning_rate": 2.3468221220716844e-05, + "loss": 2.8050756454467773, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.02831537462770939, + "learning_rate": 2.33782560345438e-05, + "loss": 2.8161096572875977, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.028237098827958107, + "learning_rate": 2.328851613455932e-05, + "loss": 2.7897520065307617, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.027944844216108322, + "learning_rate": 2.3199001593065204e-05, + "loss": 2.816229820251465, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.028575003147125244, + "learning_rate": 2.3109712482181684e-05, + "loss": 2.8069968223571777, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.028569547459483147, + "learning_rate": 2.3020648873847414e-05, + "loss": 2.78981614112854, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.028383128345012665, + "learning_rate": 2.2931810839819235e-05, + "loss": 2.820969343185425, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.027872422710061073, + "learning_rate": 2.284319845167253e-05, + "loss": 2.781301498413086, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.028342951089143753, + "learning_rate": 2.2754811780800532e-05, + "loss": 2.816469192504883, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.028373923152685165, + "learning_rate": 2.2666650898414827e-05, + "loss": 2.799290657043457, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.02807556465268135, + "learning_rate": 2.2578715875545095e-05, + "loss": 2.814924716949463, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.02789521962404251, + "learning_rate": 2.2491006783038976e-05, + "loss": 2.7950439453125, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.028297580778598785, + "learning_rate": 2.2403523691562092e-05, + "loss": 2.8535611629486084, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.02745051123201847, + "learning_rate": 2.2316266671598003e-05, + "loss": 2.7812037467956543, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.027246123179793358, + "learning_rate": 2.222923579344805e-05, + "loss": 2.778583288192749, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.027801746502518654, + "learning_rate": 2.2142431127231502e-05, + "loss": 2.812328815460205, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.027761677280068398, + "learning_rate": 2.2055852742885322e-05, + "loss": 2.829758882522583, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.028340548276901245, + "learning_rate": 2.196950071016409e-05, + "loss": 2.8203043937683105, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.027355236932635307, + "learning_rate": 2.1883375098640134e-05, + "loss": 2.777925968170166, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027668653056025505, + "learning_rate": 2.1797475977703352e-05, + "loss": 2.789498805999756, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02823868952691555, + "learning_rate": 2.1711803416560964e-05, + "loss": 2.7765135765075684, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.02812831476330757, + "learning_rate": 2.16263574842379e-05, + "loss": 2.774756669998169, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.027789659798145294, + "learning_rate": 2.15411382495764e-05, + "loss": 2.8317346572875977, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.035616710782051086, + "learning_rate": 2.1456145781236057e-05, + "loss": 2.8123791217803955, + "step": 3240 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.3813260125995008e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-3240/training_args.bin b/runs/l2r90-baseline/checkpoint-3240/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3240/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-360/chat_template.jinja b/runs/l2r90-baseline/checkpoint-360/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-360/config.json b/runs/l2r90-baseline/checkpoint-360/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-360/generation_config.json b/runs/l2r90-baseline/checkpoint-360/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-360/model.safetensors b/runs/l2r90-baseline/checkpoint-360/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b2093bf098854f2aa2361e88b041436e51ed2c95 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5518c6563c8ab5204126d1b92837f3a9273693a6812f1319d2fa9d9baff16348 +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-360/optimizer.pt b/runs/l2r90-baseline/checkpoint-360/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5741cbf54e5f61b8adb0ae9c46f767851689c892 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fac2b4effb5f939d5027f6ada65805f1279975071f7d418ee92347c24b3fa06d +size 1166825338 diff --git a/runs/l2r90-baseline/checkpoint-360/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-360/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..c3dcf55387a71ee2a6afda87f94800657089f6cc --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5cf3bac380addaa516f941e6a276bb95c92ecfb3c08371ff9fba4b2faac6df0 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-360/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-360/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..26b693a57bc20ed89f333203ee3df5b86d2cc75b --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:219db464e75bbf2b6c41a71470b6182f4de4159f65fd71e3323fd50c658eea3a +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-360/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-360/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..9f43ebb920eaa7f2483f205604373ffec78981af --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03b8316374c36bf601b0d74a53807a0d460d535e42771ed5f0f6db088f66ab63 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-360/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-360/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..c89f029f8f7c32ec2bc7c6471885e82ebfe5322e --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d6c7bdfadb70cba54ce8abee8b9fbcd1d5870419c1ccbd2e7362a47c09ba74e +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-360/scheduler.pt b/runs/l2r90-baseline/checkpoint-360/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..d63b22e047ab13a6c29adc08f78969e1d6e569ff --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2802ceffe363a15cca2d2192020ec6c680a5afa8cc5d08bb488000d79d1a8eb0 +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-360/tokenizer.json b/runs/l2r90-baseline/checkpoint-360/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-360/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-360/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-360/trainer_state.json b/runs/l2r90-baseline/checkpoint-360/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..8632b646315bb3cfa011301c5d864aba8e47cffb --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/trainer_state.json @@ -0,0 +1,2586 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.0, + "eval_steps": 500, + "global_step": 360, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.534806680666112e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-360/training_args.bin b/runs/l2r90-baseline/checkpoint-360/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-360/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-3600/chat_template.jinja b/runs/l2r90-baseline/checkpoint-3600/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-3600/config.json b/runs/l2r90-baseline/checkpoint-3600/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-3600/generation_config.json b/runs/l2r90-baseline/checkpoint-3600/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-3600/model.safetensors b/runs/l2r90-baseline/checkpoint-3600/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8744a97f301b6417af7c4e09ffb07028f35e2f44 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b0079515fa8e03d4ede34fbbc1c4a85937b053aa209c5fceaf8a4b94716babe +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-3600/optimizer.pt b/runs/l2r90-baseline/checkpoint-3600/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..50dc69f9064de595987eb41f1c4d7c2b98002fb1 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8f6d37ab699c0adaaf7cf682e015df8719cbdc2d1af74aab7e49969edb61382 +size 1166827898 diff --git a/runs/l2r90-baseline/checkpoint-3600/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-3600/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..1a514451b6a50ad9b6de0b7fc4a965cb83108e58 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6ec8d61675b917843283ea16decabe601854cecb62727b8d6fd0736b1d8e982 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-3600/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-3600/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..4b5687b836d8e7154f2f48e4093e48f7463e53d0 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf4a2454d9040e2fdfe5d3ef7263fb3bdbcaa970303baa7413c80aed2f694675 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-3600/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-3600/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..31f9a655287c5cfc3f33b435d1688e36b4b32220 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e25cf1e6918353e4d123a8988a16033aeb3a27afe509656b117bb952c8698b33 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-3600/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-3600/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..276e9dddc44b2549c7e379c8e61b12187f2dd7e3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2366911094763ba47ae31b700e5549017c3b44f5913088fc81788f708e2fa6b0 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-3600/scheduler.pt b/runs/l2r90-baseline/checkpoint-3600/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..dff04aa3b724ae7dec72031a9c3d0ecf0b695f1b --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:187927afe10d5874e052562039160b9509612fe972489c80a97620ccae701300 +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-3600/tokenizer.json b/runs/l2r90-baseline/checkpoint-3600/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-3600/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-3600/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-3600/trainer_state.json b/runs/l2r90-baseline/checkpoint-3600/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..3af15bc60ef2d14f943069807fe41aff3afe2684 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/trainer_state.json @@ -0,0 +1,25618 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 50.0, + "eval_steps": 500, + "global_step": 3600, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 5.549652099609375, + "eval_runtime": 57.761, + "eval_samples_per_second": 42.278, + "eval_steps_per_second": 1.333, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.03958690166473389, + "learning_rate": 0.0005919487226162231, + "loss": 5.459286689758301, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.03299890458583832, + "learning_rate": 0.0005918869546699877, + "loss": 5.45464563369751, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.028040863573551178, + "learning_rate": 0.0005918249539727509, + "loss": 5.444796562194824, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.028365720063447952, + "learning_rate": 0.0005917627205744655, + "loss": 5.447852611541748, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.02822224237024784, + "learning_rate": 0.0005917002545252719, + "loss": 5.453883171081543, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.036075398325920105, + "learning_rate": 0.0005916375558754977, + "loss": 5.462671279907227, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04683876410126686, + "learning_rate": 0.0005915746246756581, + "loss": 5.443917274475098, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.05457588657736778, + "learning_rate": 0.0005915114609764558, + "loss": 5.444374084472656, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.04250836744904518, + "learning_rate": 0.0005914480648287804, + "loss": 5.435029983520508, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.029477844014763832, + "learning_rate": 0.0005913844362837093, + "loss": 5.421862602233887, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.03673284500837326, + "learning_rate": 0.0005913205753925066, + "loss": 5.430391311645508, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.030489861965179443, + "learning_rate": 0.0005912564822066241, + "loss": 5.389314651489258, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.04037046805024147, + "learning_rate": 0.0005911921567777005, + "loss": 5.425060272216797, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.0445217490196228, + "learning_rate": 0.0005911275991575613, + "loss": 5.437565326690674, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.03885014355182648, + "learning_rate": 0.0005910628093982198, + "loss": 5.410930633544922, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.0334964245557785, + "learning_rate": 0.0005909977875518759, + "loss": 5.412092208862305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.03499286621809006, + "learning_rate": 0.0005909325336709164, + "loss": 5.382216453552246, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.04171644523739815, + "learning_rate": 0.0005908670478079152, + "loss": 5.387385368347168, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.057558078318834305, + "learning_rate": 0.000590801330015633, + "loss": 5.437925815582275, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.053846780210733414, + "learning_rate": 0.0005907353803470177, + "loss": 5.41644811630249, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.027259036898612976, + "learning_rate": 0.0005906691988552034, + "loss": 5.409185409545898, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.03394043818116188, + "learning_rate": 0.0005906027855935115, + "loss": 5.377973556518555, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.03731574863195419, + "learning_rate": 0.0005905361406154501, + "loss": 5.382505893707275, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.04705141484737396, + "learning_rate": 0.0005904692639747137, + "loss": 5.413661003112793, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.044811323285102844, + "learning_rate": 0.0005904021557251837, + "loss": 5.417835235595703, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.04000752419233322, + "learning_rate": 0.0005903348159209277, + "loss": 5.406160354614258, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.038364164531230927, + "learning_rate": 0.0005902672446162007, + "loss": 5.357950687408447, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.030500127002596855, + "learning_rate": 0.0005901994418654432, + "loss": 5.344951629638672, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.030644729733467102, + "learning_rate": 0.0005901314077232829, + "loss": 5.3599348068237305, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.037976790219545364, + "learning_rate": 0.0005900631422445335, + "loss": 5.354084491729736, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.04106011986732483, + "learning_rate": 0.0005899946454841955, + "loss": 5.363678455352783, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.04334832355380058, + "learning_rate": 0.0005899259174974553, + "loss": 5.350959300994873, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.049939028918743134, + "learning_rate": 0.000589856958339686, + "loss": 5.3317155838012695, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.05300058051943779, + "learning_rate": 0.0005897877680664465, + "loss": 5.353314399719238, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.05943184345960617, + "learning_rate": 0.0005897183467334822, + "loss": 5.3683929443359375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.05600789934396744, + "learning_rate": 0.0005896486943967249, + "loss": 5.345516204833984, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.05392718315124512, + "learning_rate": 0.0005895788111122919, + "loss": 5.291389465332031, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.037913475185632706, + "learning_rate": 0.0005895086969364871, + "loss": 5.353423118591309, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.03362123295664787, + "learning_rate": 0.0005894383519258001, + "loss": 5.379500389099121, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032558735460042953, + "learning_rate": 0.0005893677761369066, + "loss": 5.34043025970459, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.04074402526021004, + "learning_rate": 0.0005892969696266683, + "loss": 5.303869247436523, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.04270484298467636, + "learning_rate": 0.0005892259324521328, + "loss": 5.27297830581665, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.04604959115386009, + "learning_rate": 0.0005891546646705334, + "loss": 5.313268184661865, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.04462246224284172, + "learning_rate": 0.0005890831663392891, + "loss": 5.31167459487915, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.04966094344854355, + "learning_rate": 0.0005890114375160052, + "loss": 5.316695690155029, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.04662073776125908, + "learning_rate": 0.0005889394782584718, + "loss": 5.316898822784424, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.04380315542221069, + "learning_rate": 0.0005888672886246656, + "loss": 5.285087585449219, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03239579498767853, + "learning_rate": 0.0005887948686727483, + "loss": 5.304283142089844, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.029690183699131012, + "learning_rate": 0.0005887222184610675, + "loss": 5.305715084075928, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.03444967418909073, + "learning_rate": 0.0005886493380481559, + "loss": 5.2707672119140625, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.03302565962076187, + "learning_rate": 0.0005885762274927322, + "loss": 5.277251243591309, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.028477763757109642, + "learning_rate": 0.0005885028868537, + "loss": 5.295318603515625, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.022988801822066307, + "learning_rate": 0.0005884293161901487, + "loss": 5.249057292938232, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.026115216314792633, + "learning_rate": 0.000588355515561353, + "loss": 5.270636081695557, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.02576332725584507, + "learning_rate": 0.0005882814850267723, + "loss": 5.263096332550049, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.023978352546691895, + "learning_rate": 0.0005882072246460521, + "loss": 5.276605606079102, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.020695513114333153, + "learning_rate": 0.0005881327344790223, + "loss": 5.2141432762146, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01952219195663929, + "learning_rate": 0.0005880580145856986, + "loss": 5.214289665222168, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.022633206099271774, + "learning_rate": 0.0005879830650262813, + "loss": 5.236690998077393, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.023082347586750984, + "learning_rate": 0.0005879078858611557, + "loss": 5.2092156410217285, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.019620463252067566, + "learning_rate": 0.0005878324771508927, + "loss": 5.264451503753662, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.021504493430256844, + "learning_rate": 0.0005877568389562475, + "loss": 5.23444938659668, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.025672364979982376, + "learning_rate": 0.0005876809713381606, + "loss": 5.205289363861084, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.034436631947755814, + "learning_rate": 0.0005876048743577569, + "loss": 5.2047648429870605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.042389530688524246, + "learning_rate": 0.0005875285480763466, + "loss": 5.192008018493652, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.03310716897249222, + "learning_rate": 0.0005874519925554244, + "loss": 5.162734031677246, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.03029288910329342, + "learning_rate": 0.0005873752078566694, + "loss": 5.212151527404785, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04619310051202774, + "learning_rate": 0.0005872981940419459, + "loss": 5.215909004211426, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.05618087574839592, + "learning_rate": 0.0005872209511733027, + "loss": 5.210192680358887, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.03783707693219185, + "learning_rate": 0.0005871434793129726, + "loss": 5.201597690582275, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.03721104562282562, + "learning_rate": 0.0005870657785233735, + "loss": 5.188891410827637, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.033336155116558075, + "learning_rate": 0.0005869878488671075, + "loss": 5.1598615646362305, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.2592363357543945, + "eval_runtime": 57.3343, + "eval_samples_per_second": 42.592, + "eval_steps_per_second": 1.343, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03719741478562355, + "learning_rate": 0.0005869096904069611, + "loss": 5.1646623611450195, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.04888090118765831, + "learning_rate": 0.0005868313032059052, + "loss": 5.171182155609131, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.05073149502277374, + "learning_rate": 0.0005867526873270949, + "loss": 5.183949947357178, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.04187731072306633, + "learning_rate": 0.0005866738428338695, + "loss": 5.176133155822754, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.039445552974939346, + "learning_rate": 0.000586594769789753, + "loss": 5.148100852966309, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.035551708191633224, + "learning_rate": 0.0005865154682584524, + "loss": 5.151512145996094, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.031175289303064346, + "learning_rate": 0.0005864359383038602, + "loss": 5.102468967437744, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.03696789592504501, + "learning_rate": 0.000586356179990052, + "loss": 5.119902610778809, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.04836135357618332, + "learning_rate": 0.0005862761933812875, + "loss": 5.15183162689209, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.05206914246082306, + "learning_rate": 0.0005861959785420106, + "loss": 5.154870986938477, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.046368084847927094, + "learning_rate": 0.000586115535536849, + "loss": 5.130651473999023, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.07146432995796204, + "learning_rate": 0.000586034864430614, + "loss": 5.175804138183594, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.08809789270162582, + "learning_rate": 0.0005859539652883008, + "loss": 5.229716777801514, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.08116360008716583, + "learning_rate": 0.0005858728381750888, + "loss": 5.21910285949707, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.08282570540904999, + "learning_rate": 0.0005857914831563402, + "loss": 5.2336506843566895, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.07403547316789627, + "learning_rate": 0.0005857099002976016, + "loss": 5.251604080200195, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.07024004310369492, + "learning_rate": 0.0005856280896646026, + "loss": 5.267114162445068, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.056330062448978424, + "learning_rate": 0.0005855460513232567, + "loss": 5.201976299285889, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.051763955503702164, + "learning_rate": 0.0005854637853396606, + "loss": 5.1733198165893555, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.039900992065668106, + "learning_rate": 0.0005853812917800945, + "loss": 5.222941875457764, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.04187863692641258, + "learning_rate": 0.0005852985707110221, + "loss": 5.165033340454102, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.042645812034606934, + "learning_rate": 0.0005852156221990901, + "loss": 5.157554626464844, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.03856075555086136, + "learning_rate": 0.0005851324463111289, + "loss": 5.184629440307617, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.029387330636382103, + "learning_rate": 0.0005850490431141516, + "loss": 5.118739128112793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.027584781870245934, + "learning_rate": 0.0005849654126753545, + "loss": 5.100855350494385, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.023655924946069717, + "learning_rate": 0.0005848815550621175, + "loss": 5.154520034790039, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.024163050577044487, + "learning_rate": 0.0005847974703420028, + "loss": 5.105189323425293, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.020369920879602432, + "learning_rate": 0.0005847131585827564, + "loss": 5.113671779632568, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.021030737087130547, + "learning_rate": 0.0005846286198523061, + "loss": 5.0847649574279785, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.01729249767959118, + "learning_rate": 0.0005845438542187638, + "loss": 5.053225517272949, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.018704712390899658, + "learning_rate": 0.0005844588617504236, + "loss": 5.079106330871582, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.02049064077436924, + "learning_rate": 0.0005843736425157621, + "loss": 5.044185161590576, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.020736856386065483, + "learning_rate": 0.000584288196583439, + "loss": 5.041932106018066, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.016543276607990265, + "learning_rate": 0.0005842025240222966, + "loss": 5.032522201538086, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.018414106220006943, + "learning_rate": 0.0005841166249013598, + "loss": 5.037397384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.01611698977649212, + "learning_rate": 0.0005840304992898359, + "loss": 5.038179397583008, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.014310561120510101, + "learning_rate": 0.0005839441472571147, + "loss": 5.0146484375, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.015472214668989182, + "learning_rate": 0.0005838575688727685, + "loss": 5.060323715209961, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.016373401507735252, + "learning_rate": 0.0005837707642065518, + "loss": 5.032100677490234, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.018694665282964706, + "learning_rate": 0.000583683733328402, + "loss": 5.036993026733398, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.019977087154984474, + "learning_rate": 0.0005835964763084378, + "loss": 4.991037368774414, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.023741086944937706, + "learning_rate": 0.000583508993216961, + "loss": 5.043832778930664, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.025061562657356262, + "learning_rate": 0.0005834212841244548, + "loss": 5.016516208648682, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.020848384127020836, + "learning_rate": 0.0005833333491015851, + "loss": 4.986289978027344, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.015086354687809944, + "learning_rate": 0.0005832451882191995, + "loss": 4.9948625564575195, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.015096964314579964, + "learning_rate": 0.0005831568015483274, + "loss": 4.996342658996582, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016932200640439987, + "learning_rate": 0.0005830681891601807, + "loss": 4.998758316040039, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.019936930388212204, + "learning_rate": 0.0005829793511261525, + "loss": 4.959854602813721, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.027658682316541672, + "learning_rate": 0.0005828902875178183, + "loss": 4.968629837036133, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.03299275413155556, + "learning_rate": 0.0005828009984069347, + "loss": 4.994814872741699, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.033717650920152664, + "learning_rate": 0.0005827114838654406, + "loss": 4.976295471191406, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.02969924360513687, + "learning_rate": 0.0005826217439654562, + "loss": 4.969155788421631, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.02846827358007431, + "learning_rate": 0.0005825317787792831, + "loss": 4.989721298217773, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.03670600801706314, + "learning_rate": 0.0005824415883794049, + "loss": 4.988689422607422, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.045570675283670425, + "learning_rate": 0.0005823511728384863, + "loss": 4.974407196044922, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.04564111679792404, + "learning_rate": 0.0005822605322293733, + "loss": 4.966331481933594, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.04173210263252258, + "learning_rate": 0.0005821696666250937, + "loss": 4.925197601318359, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.04228242486715317, + "learning_rate": 0.0005820785760988559, + "loss": 4.965020656585693, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.05144782364368439, + "learning_rate": 0.0005819872607240503, + "loss": 5.0022149085998535, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.049400657415390015, + "learning_rate": 0.0005818957205742478, + "loss": 4.9635114669799805, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.043692708015441895, + "learning_rate": 0.0005818039557232005, + "loss": 4.977723121643066, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.04891742020845413, + "learning_rate": 0.0005817119662448421, + "loss": 4.996312141418457, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.035564981400966644, + "learning_rate": 0.0005816197522132866, + "loss": 4.940932273864746, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04679827764630318, + "learning_rate": 0.0005815273137028292, + "loss": 4.983131408691406, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.05489890277385712, + "learning_rate": 0.0005814346507879459, + "loss": 4.977827548980713, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04640809819102287, + "learning_rate": 0.0005813417635432937, + "loss": 4.954761505126953, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04439452290534973, + "learning_rate": 0.00058124865204371, + "loss": 4.914865970611572, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.04587096720933914, + "learning_rate": 0.0005811553163642131, + "loss": 4.949592590332031, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.044119734317064285, + "learning_rate": 0.0005810617565800019, + "loss": 4.956068992614746, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.043133001774549484, + "learning_rate": 0.0005809679727664559, + "loss": 4.993522644042969, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04315278306603432, + "learning_rate": 0.0005808739649991348, + "loss": 4.894291877746582, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04819472134113312, + "learning_rate": 0.0005807797333537792, + "loss": 4.958306789398193, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.023706436157227, + "eval_runtime": 57.4082, + "eval_samples_per_second": 42.537, + "eval_steps_per_second": 1.341, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.04403403028845787, + "learning_rate": 0.0005806852779063098, + "loss": 4.884971618652344, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.041993748396635056, + "learning_rate": 0.0005805905987328275, + "loss": 4.916353225708008, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.042302004992961884, + "learning_rate": 0.0005804956959096138, + "loss": 4.9012556076049805, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.044500596821308136, + "learning_rate": 0.0005804005695131299, + "loss": 4.915240287780762, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.03914736211299896, + "learning_rate": 0.0005803052196200177, + "loss": 4.885915756225586, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.0361703559756279, + "learning_rate": 0.0005802096463070988, + "loss": 4.829287528991699, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03432302549481392, + "learning_rate": 0.0005801138496513749, + "loss": 4.926834583282471, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.0317804180085659, + "learning_rate": 0.0005800178297300276, + "loss": 4.8707170486450195, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.035261351615190506, + "learning_rate": 0.0005799215866204183, + "loss": 4.897791862487793, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.03276067599654198, + "learning_rate": 0.0005798251204000886, + "loss": 4.864504814147949, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.034757066518068314, + "learning_rate": 0.0005797284311467594, + "loss": 4.890812873840332, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.033726081252098083, + "learning_rate": 0.0005796315189383316, + "loss": 4.831684112548828, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.03074726276099682, + "learning_rate": 0.0005795343838528855, + "loss": 4.8534111976623535, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02215893566608429, + "learning_rate": 0.0005794370259686811, + "loss": 4.8311238288879395, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.02132033370435238, + "learning_rate": 0.0005793394453641581, + "loss": 4.842617034912109, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020825816318392754, + "learning_rate": 0.0005792416421179351, + "loss": 4.84535026550293, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.019334357231855392, + "learning_rate": 0.0005791436163088106, + "loss": 4.839937686920166, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017372334375977516, + "learning_rate": 0.0005790453680157622, + "loss": 4.8115553855896, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.0190406683832407, + "learning_rate": 0.0005789468973179468, + "loss": 4.832542419433594, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.017372574657201767, + "learning_rate": 0.0005788482042947004, + "loss": 4.793632507324219, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.019171632826328278, + "learning_rate": 0.0005787492890255382, + "loss": 4.776987552642822, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.02109278179705143, + "learning_rate": 0.0005786501515901545, + "loss": 4.8216657638549805, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.026234494522213936, + "learning_rate": 0.0005785507920684225, + "loss": 4.8109049797058105, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.0357641875743866, + "learning_rate": 0.0005784512105403944, + "loss": 4.821109771728516, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.03627947345376015, + "learning_rate": 0.000578351407086301, + "loss": 4.810632228851318, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.032253146171569824, + "learning_rate": 0.0005782513817865527, + "loss": 4.7930521965026855, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.03402210772037506, + "learning_rate": 0.0005781511347217375, + "loss": 4.772500991821289, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.0321584977209568, + "learning_rate": 0.000578050665972623, + "loss": 4.781399726867676, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.030242707580327988, + "learning_rate": 0.0005779499756201549, + "loss": 4.802542686462402, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.02361336164176464, + "learning_rate": 0.0005778490637454576, + "loss": 4.771912574768066, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.028781302273273468, + "learning_rate": 0.000577747930429834, + "loss": 4.796924591064453, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.027042683213949203, + "learning_rate": 0.0005776465757547654, + "loss": 4.745218276977539, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.02715984545648098, + "learning_rate": 0.0005775449998019113, + "loss": 4.792944431304932, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.0314096100628376, + "learning_rate": 0.00057744320265311, + "loss": 4.7649993896484375, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.028749193996191025, + "learning_rate": 0.0005773411843903773, + "loss": 4.7980546951293945, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.02655309997498989, + "learning_rate": 0.0005772389450959075, + "loss": 4.746520519256592, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.027497397735714912, + "learning_rate": 0.000577136484852073, + "loss": 4.7490715980529785, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.0343393050134182, + "learning_rate": 0.000577033803741424, + "loss": 4.74436092376709, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.04444412514567375, + "learning_rate": 0.0005769309018466891, + "loss": 4.755973815917969, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.0437556728720665, + "learning_rate": 0.0005768277792507744, + "loss": 4.784368515014648, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0369921550154686, + "learning_rate": 0.0005767244360367638, + "loss": 4.753410339355469, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.05038141459226608, + "learning_rate": 0.0005766208722879192, + "loss": 4.755491256713867, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.04511422663927078, + "learning_rate": 0.00057651708808768, + "loss": 4.760293006896973, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.048762883991003036, + "learning_rate": 0.0005764130835196632, + "loss": 4.755958080291748, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.06749686598777771, + "learning_rate": 0.0005763088586676634, + "loss": 4.802557468414307, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.06451250612735748, + "learning_rate": 0.0005762044136156528, + "loss": 4.812925338745117, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.05436156690120697, + "learning_rate": 0.0005760997484477809, + "loss": 4.807103157043457, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.05233754590153694, + "learning_rate": 0.0005759948632483742, + "loss": 4.782449722290039, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.05519746616482735, + "learning_rate": 0.0005758897581019371, + "loss": 4.808449745178223, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.0578567236661911, + "learning_rate": 0.000575784433093151, + "loss": 4.7676005363464355, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.0588264986872673, + "learning_rate": 0.0005756788883068743, + "loss": 4.805377960205078, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04225790873169899, + "learning_rate": 0.0005755731238281423, + "loss": 4.748775005340576, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.0376368910074234, + "learning_rate": 0.0005754671397421678, + "loss": 4.759791374206543, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03830771893262863, + "learning_rate": 0.0005753609361343402, + "loss": 4.729974746704102, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.03684520348906517, + "learning_rate": 0.0005752545130902256, + "loss": 4.729854583740234, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02983408235013485, + "learning_rate": 0.0005751478706955674, + "loss": 4.78502082824707, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02854904904961586, + "learning_rate": 0.0005750410090362854, + "loss": 4.717472076416016, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.023819150403141975, + "learning_rate": 0.0005749339281984761, + "loss": 4.717124938964844, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.02477158047258854, + "learning_rate": 0.0005748266282684124, + "loss": 4.71781063079834, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.022002534940838814, + "learning_rate": 0.0005747191093325443, + "loss": 4.7480669021606445, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.02133483812212944, + "learning_rate": 0.0005746113714774976, + "loss": 4.693770408630371, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.022407138720154762, + "learning_rate": 0.0005745034147900747, + "loss": 4.720734119415283, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.020850926637649536, + "learning_rate": 0.0005743952393572544, + "loss": 4.745505332946777, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.020420582965016365, + "learning_rate": 0.0005742868452661918, + "loss": 4.688610076904297, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.019014107063412666, + "learning_rate": 0.0005741782326042181, + "loss": 4.673436164855957, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.017393039539456367, + "learning_rate": 0.0005740694014588403, + "loss": 4.6941118240356445, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.017877329140901566, + "learning_rate": 0.0005739603519177419, + "loss": 4.667474269866943, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.017749635502696037, + "learning_rate": 0.0005738510840687821, + "loss": 4.6901092529296875, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.018746376037597656, + "learning_rate": 0.000573741597999996, + "loss": 4.680166721343994, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016260217875242233, + "learning_rate": 0.0005736318937995947, + "loss": 4.6799821853637695, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015099814161658287, + "learning_rate": 0.0005735219715559646, + "loss": 4.680666923522949, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.016539735719561577, + "learning_rate": 0.0005734118313576683, + "loss": 4.639509201049805, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 4.768754482269287, + "eval_runtime": 58.3352, + "eval_samples_per_second": 41.862, + "eval_steps_per_second": 1.32, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.019733518362045288, + "learning_rate": 0.0005733014732934436, + "loss": 4.635312080383301, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.022571036592125893, + "learning_rate": 0.0005731908974522042, + "loss": 4.6041765213012695, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.022238552570343018, + "learning_rate": 0.0005730801039230389, + "loss": 4.648976802825928, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.02115180902183056, + "learning_rate": 0.0005729690927952123, + "loss": 4.623968124389648, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.021789928898215294, + "learning_rate": 0.0005728578641581637, + "loss": 4.620028018951416, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.02593887411057949, + "learning_rate": 0.0005727464181015081, + "loss": 4.591062545776367, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.0347842313349247, + "learning_rate": 0.0005726347547150357, + "loss": 4.632149696350098, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04119519516825676, + "learning_rate": 0.0005725228740887117, + "loss": 4.6254191398620605, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.04895750433206558, + "learning_rate": 0.0005724107763126761, + "loss": 4.621372222900391, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.05752161890268326, + "learning_rate": 0.0005722984614772442, + "loss": 4.61965274810791, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05115870013833046, + "learning_rate": 0.000572185929672906, + "loss": 4.63119649887085, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.05721089243888855, + "learning_rate": 0.0005720731809903263, + "loss": 4.652647018432617, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.0646839588880539, + "learning_rate": 0.000571960215520345, + "loss": 4.642184257507324, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.0682854950428009, + "learning_rate": 0.0005718470333539757, + "loss": 4.676589012145996, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.057402849197387695, + "learning_rate": 0.0005717336345824077, + "loss": 4.6585283279418945, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.05885373800992966, + "learning_rate": 0.0005716200192970043, + "loss": 4.671728134155273, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.05265705659985542, + "learning_rate": 0.0005715061875893032, + "loss": 4.641249656677246, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.05417842045426369, + "learning_rate": 0.0005713921395510166, + "loss": 4.661381721496582, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.06602369248867035, + "learning_rate": 0.0005712778752740307, + "loss": 4.684137344360352, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.06004182621836662, + "learning_rate": 0.0005711633948504066, + "loss": 4.625332832336426, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.04801511764526367, + "learning_rate": 0.0005710486983723787, + "loss": 4.664791107177734, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.04380105435848236, + "learning_rate": 0.0005709337859323561, + "loss": 4.664966583251953, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.03858727216720581, + "learning_rate": 0.0005708186576229218, + "loss": 4.661867618560791, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03240576758980751, + "learning_rate": 0.0005707033135368323, + "loss": 4.5855326652526855, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.028561925515532494, + "learning_rate": 0.0005705877537670184, + "loss": 4.637331008911133, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.026389246806502342, + "learning_rate": 0.0005704719784065846, + "loss": 4.598666667938232, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.02484819106757641, + "learning_rate": 0.0005703559875488088, + "loss": 4.601729393005371, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.02522977627813816, + "learning_rate": 0.0005702397812871429, + "loss": 4.609833717346191, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02589508146047592, + "learning_rate": 0.0005701233597152121, + "loss": 4.574592590332031, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.023374728858470917, + "learning_rate": 0.0005700067229268154, + "loss": 4.61362361907959, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.02465851791203022, + "learning_rate": 0.0005698898710159245, + "loss": 4.639899253845215, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.02258715033531189, + "learning_rate": 0.0005697728040766852, + "loss": 4.5656538009643555, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.018009789288043976, + "learning_rate": 0.0005696555222034162, + "loss": 4.61449670791626, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02121281623840332, + "learning_rate": 0.0005695380254906094, + "loss": 4.553830623626709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.018078701570630074, + "learning_rate": 0.0005694203140329296, + "loss": 4.534579277038574, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.0169303547590971, + "learning_rate": 0.000569302387925215, + "loss": 4.533995628356934, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.016318142414093018, + "learning_rate": 0.0005691842472624764, + "loss": 4.544438362121582, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.016664575785398483, + "learning_rate": 0.0005690658921398977, + "loss": 4.5913286209106445, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.01641504280269146, + "learning_rate": 0.0005689473226528354, + "loss": 4.520743370056152, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.017012735828757286, + "learning_rate": 0.0005688285388968187, + "loss": 4.517453193664551, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01672312617301941, + "learning_rate": 0.0005687095409675499, + "loss": 4.5235915184021, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.016893865540623665, + "learning_rate": 0.000568590328960903, + "loss": 4.543200492858887, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.017530135810375214, + "learning_rate": 0.0005684709029729253, + "loss": 4.54912805557251, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.018698599189519882, + "learning_rate": 0.0005683512630998361, + "loss": 4.522650718688965, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01984071172773838, + "learning_rate": 0.000568231409438027, + "loss": 4.546173095703125, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.023218289017677307, + "learning_rate": 0.0005681113420840619, + "loss": 4.5305094718933105, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.028968270868062973, + "learning_rate": 0.000567991061134677, + "loss": 4.521440505981445, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.032370273023843765, + "learning_rate": 0.0005678705666867805, + "loss": 4.54168701171875, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.033967748284339905, + "learning_rate": 0.0005677498588374524, + "loss": 4.538717269897461, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.028757929801940918, + "learning_rate": 0.0005676289376839452, + "loss": 4.517228603363037, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.029757825657725334, + "learning_rate": 0.0005675078033236827, + "loss": 4.510310173034668, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.03581003099679947, + "learning_rate": 0.0005673864558542605, + "loss": 4.506901264190674, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.03483985736966133, + "learning_rate": 0.0005672648953734462, + "loss": 4.549284934997559, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.031016027554869652, + "learning_rate": 0.000567143121979179, + "loss": 4.530084609985352, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.029373526573181152, + "learning_rate": 0.0005670211357695693, + "loss": 4.492218971252441, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.031661372631788254, + "learning_rate": 0.0005668989368428994, + "loss": 4.502824783325195, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.027251387014985085, + "learning_rate": 0.0005667765252976227, + "loss": 4.49193811416626, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02263176068663597, + "learning_rate": 0.0005666539012323639, + "loss": 4.512019157409668, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.02355228364467621, + "learning_rate": 0.0005665310647459189, + "loss": 4.508917331695557, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.023297371342778206, + "learning_rate": 0.0005664080159372551, + "loss": 4.497771263122559, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.019364114850759506, + "learning_rate": 0.0005662847549055107, + "loss": 4.51470947265625, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.023899901658296585, + "learning_rate": 0.0005661612817499947, + "loss": 4.48771858215332, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.02330983616411686, + "learning_rate": 0.0005660375965701872, + "loss": 4.478496551513672, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.027252234518527985, + "learning_rate": 0.0005659136994657392, + "loss": 4.497672080993652, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.030886363238096237, + "learning_rate": 0.0005657895905364723, + "loss": 4.49789571762085, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02515559457242489, + "learning_rate": 0.0005656652698823788, + "loss": 4.490771293640137, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020107125863432884, + "learning_rate": 0.0005655407376036217, + "loss": 4.481879234313965, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.022076716646552086, + "learning_rate": 0.0005654159938005345, + "loss": 4.494697570800781, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.025525815784931183, + "learning_rate": 0.0005652910385736207, + "loss": 4.517453193664551, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.03005894646048546, + "learning_rate": 0.0005651658720235548, + "loss": 4.453815460205078, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.030178502202033997, + "learning_rate": 0.0005650404942511812, + "loss": 4.493922233581543, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.029249394312500954, + "learning_rate": 0.0005649149053575141, + "loss": 4.453763008117676, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 4.607339382171631, + "eval_runtime": 57.9216, + "eval_samples_per_second": 42.16, + "eval_steps_per_second": 1.329, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.03362950682640076, + "learning_rate": 0.0005647891054437386, + "loss": 4.430700302124023, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.04223684221506119, + "learning_rate": 0.0005646630946112096, + "loss": 4.486915588378906, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.05192171037197113, + "learning_rate": 0.0005645368729614512, + "loss": 4.467541694641113, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.042525868862867355, + "learning_rate": 0.0005644104405961582, + "loss": 4.45252799987793, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.0406370609998703, + "learning_rate": 0.0005642837976171949, + "loss": 4.414548873901367, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.044476233422756195, + "learning_rate": 0.0005641569441265952, + "loss": 4.453492164611816, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.04687326401472092, + "learning_rate": 0.0005640298802265626, + "loss": 4.477725028991699, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.04011611267924309, + "learning_rate": 0.0005639026060194704, + "loss": 4.423356533050537, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.04971884936094284, + "learning_rate": 0.0005637751216078609, + "loss": 4.4853515625, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.05863234028220177, + "learning_rate": 0.0005636474270944461, + "loss": 4.466001033782959, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.057710036635398865, + "learning_rate": 0.0005635195225821072, + "loss": 4.466723442077637, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.05155074968934059, + "learning_rate": 0.0005633914081738945, + "loss": 4.4823198318481445, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.04535634070634842, + "learning_rate": 0.0005632630839730275, + "loss": 4.471219062805176, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.045147854834795, + "learning_rate": 0.0005631345500828946, + "loss": 4.476888179779053, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.03730619326233864, + "learning_rate": 0.0005630058066070533, + "loss": 4.428932189941406, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.034599728882312775, + "learning_rate": 0.0005628768536492299, + "loss": 4.47627592086792, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.032904356718063354, + "learning_rate": 0.0005627476913133193, + "loss": 4.422671794891357, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.03356337174773216, + "learning_rate": 0.0005626183197033856, + "loss": 4.456406593322754, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.035042595118284225, + "learning_rate": 0.0005624887389236609, + "loss": 4.371820449829102, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.03269032761454582, + "learning_rate": 0.0005623589490785462, + "loss": 4.402605056762695, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.028261031955480576, + "learning_rate": 0.0005622289502726108, + "loss": 4.433136940002441, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03301700949668884, + "learning_rate": 0.0005620987426105925, + "loss": 4.388359546661377, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.031664177775382996, + "learning_rate": 0.000561968326197397, + "loss": 4.41859245300293, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.026281699538230896, + "learning_rate": 0.0005618377011380988, + "loss": 4.406128883361816, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.023097911849617958, + "learning_rate": 0.0005617068675379398, + "loss": 4.389717102050781, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.022349286824464798, + "learning_rate": 0.0005615758255023306, + "loss": 4.408173561096191, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.021958226338028908, + "learning_rate": 0.0005614445751368491, + "loss": 4.434830188751221, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024551713839173317, + "learning_rate": 0.0005613131165472415, + "loss": 4.386944770812988, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.028048967942595482, + "learning_rate": 0.0005611814498394216, + "loss": 4.402279376983643, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02733907848596573, + "learning_rate": 0.0005610495751194708, + "loss": 4.403069496154785, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.020858559757471085, + "learning_rate": 0.0005609174924936384, + "loss": 4.406920909881592, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.021491149440407753, + "learning_rate": 0.0005607852020683409, + "loss": 4.385799407958984, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.021495329216122627, + "learning_rate": 0.0005606527039501621, + "loss": 4.4252495765686035, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.019708355888724327, + "learning_rate": 0.0005605199982458535, + "loss": 4.388864994049072, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.01886274851858616, + "learning_rate": 0.0005603870850623338, + "loss": 4.335962295532227, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.01955341547727585, + "learning_rate": 0.0005602539645066884, + "loss": 4.373560905456543, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.018962012603878975, + "learning_rate": 0.0005601206366861706, + "loss": 4.380022048950195, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.016663335263729095, + "learning_rate": 0.0005599871017081999, + "loss": 4.421034336090088, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01676873490214348, + "learning_rate": 0.0005598533596803631, + "loss": 4.382775783538818, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.014379126951098442, + "learning_rate": 0.0005597194107104137, + "loss": 4.41373872756958, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.015897583216428757, + "learning_rate": 0.000559585254906272, + "loss": 4.377284049987793, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.016513120383024216, + "learning_rate": 0.000559450892376025, + "loss": 4.41200065612793, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.016625918447971344, + "learning_rate": 0.000559316323227926, + "loss": 4.381187438964844, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.016028188169002533, + "learning_rate": 0.000559181547570395, + "loss": 4.353909015655518, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01682914048433304, + "learning_rate": 0.0005590465655120183, + "loss": 4.382326126098633, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.017078889533877373, + "learning_rate": 0.0005589113771615486, + "loss": 4.334949493408203, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.018182938918471336, + "learning_rate": 0.0005587759826279046, + "loss": 4.348130226135254, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01940886303782463, + "learning_rate": 0.0005586403820201713, + "loss": 4.339404106140137, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.020644692704081535, + "learning_rate": 0.0005585045754475995, + "loss": 4.358404159545898, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02433430217206478, + "learning_rate": 0.0005583685630196064, + "loss": 4.3660502433776855, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.027824537828564644, + "learning_rate": 0.0005582323448457744, + "loss": 4.34050178527832, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.02973582223057747, + "learning_rate": 0.0005580959210358523, + "loss": 4.348322868347168, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.030396128073334694, + "learning_rate": 0.0005579592916997542, + "loss": 4.356224060058594, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.029760530218482018, + "learning_rate": 0.0005578224569475599, + "loss": 4.357470989227295, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.02855013869702816, + "learning_rate": 0.0005576854168895147, + "loss": 4.335353374481201, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03376635164022446, + "learning_rate": 0.0005575481716360292, + "loss": 4.331794738769531, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.034431129693984985, + "learning_rate": 0.0005574107212976796, + "loss": 4.336880207061768, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.03524135425686836, + "learning_rate": 0.000557273065985207, + "loss": 4.31537389755249, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.03152058273553848, + "learning_rate": 0.0005571352058095179, + "loss": 4.37662410736084, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02878640592098236, + "learning_rate": 0.0005569971408816838, + "loss": 4.292388916015625, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.029663654044270515, + "learning_rate": 0.0005568588713129409, + "loss": 4.344346523284912, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.027895759791135788, + "learning_rate": 0.0005567203972146906, + "loss": 4.365072727203369, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.026465769857168198, + "learning_rate": 0.000556581718698499, + "loss": 4.334959506988525, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.03191852569580078, + "learning_rate": 0.000556442835876097, + "loss": 4.361069679260254, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.03535444661974907, + "learning_rate": 0.0005563037488593799, + "loss": 4.349765777587891, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03451581299304962, + "learning_rate": 0.0005561644577604074, + "loss": 4.343142509460449, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029810845851898193, + "learning_rate": 0.000556024962691404, + "loss": 4.307485580444336, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.027523217722773552, + "learning_rate": 0.0005558852637647584, + "loss": 4.366406440734863, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03043208085000515, + "learning_rate": 0.0005557453610930235, + "loss": 4.356929779052734, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.03058922104537487, + "learning_rate": 0.0005556052547889161, + "loss": 4.348942279815674, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.032052166759967804, + "learning_rate": 0.0005554649449653177, + "loss": 4.3496809005737305, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.03170694038271904, + "learning_rate": 0.0005553244317352731, + "loss": 4.315959930419922, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 4.4757890701293945, + "eval_runtime": 57.7451, + "eval_samples_per_second": 42.289, + "eval_steps_per_second": 1.333, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.029025450348854065, + "learning_rate": 0.0005551837152119915, + "loss": 4.288647174835205, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.029403962194919586, + "learning_rate": 0.0005550427955088455, + "loss": 4.288555145263672, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.03265019878745079, + "learning_rate": 0.0005549016727393715, + "loss": 4.292625904083252, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.0387202650308609, + "learning_rate": 0.0005547603470172697, + "loss": 4.238926887512207, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.04554206505417824, + "learning_rate": 0.0005546188184564036, + "loss": 4.29966402053833, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.036231476813554764, + "learning_rate": 0.0005544770871708004, + "loss": 4.304398536682129, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.038703009486198425, + "learning_rate": 0.0005543351532746501, + "loss": 4.2945356369018555, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03224619850516319, + "learning_rate": 0.0005541930168823067, + "loss": 4.261351585388184, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.029722856357693672, + "learning_rate": 0.0005540506781082864, + "loss": 4.303143501281738, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.02764045260846615, + "learning_rate": 0.0005539081370672695, + "loss": 4.243679523468018, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.0312200877815485, + "learning_rate": 0.0005537653938740984, + "loss": 4.279904365539551, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.027349065989255905, + "learning_rate": 0.0005536224486437788, + "loss": 4.270228385925293, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.030385233461856842, + "learning_rate": 0.0005534793014914793, + "loss": 4.301590442657471, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.03292274847626686, + "learning_rate": 0.0005533359525325307, + "loss": 4.301828384399414, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.03265977278351784, + "learning_rate": 0.0005531924018824266, + "loss": 4.286200523376465, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.03177836537361145, + "learning_rate": 0.0005530486496568236, + "loss": 4.296319007873535, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.028879934921860695, + "learning_rate": 0.0005529046959715397, + "loss": 4.261706829071045, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.02855812944471836, + "learning_rate": 0.0005527605409425562, + "loss": 4.273903846740723, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03037901036441326, + "learning_rate": 0.0005526161846860161, + "loss": 4.248080253601074, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.029279105365276337, + "learning_rate": 0.0005524716273182245, + "loss": 4.281311988830566, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.02592580020427704, + "learning_rate": 0.0005523268689556489, + "loss": 4.204848289489746, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.021899603307247162, + "learning_rate": 0.0005521819097149183, + "loss": 4.232453346252441, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.019704442471265793, + "learning_rate": 0.0005520367497128238, + "loss": 4.253462314605713, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.019550519064068794, + "learning_rate": 0.0005518913890663182, + "loss": 4.307737350463867, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.020320113748311996, + "learning_rate": 0.0005517458278925161, + "loss": 4.257411956787109, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02118559181690216, + "learning_rate": 0.0005516000663086932, + "loss": 4.262089729309082, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.02050616219639778, + "learning_rate": 0.0005514541044322872, + "loss": 4.2490644454956055, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019793255254626274, + "learning_rate": 0.0005513079423808967, + "loss": 4.267400741577148, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.020760254934430122, + "learning_rate": 0.0005511615802722823, + "loss": 4.2532429695129395, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.019524799659848213, + "learning_rate": 0.0005510150182243647, + "loss": 4.267422199249268, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.020166637375950813, + "learning_rate": 0.0005508682563552269, + "loss": 4.214296817779541, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.01964394748210907, + "learning_rate": 0.0005507212947831119, + "loss": 4.252087116241455, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01690433733165264, + "learning_rate": 0.0005505741336264239, + "loss": 4.232336044311523, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.017682762816548347, + "learning_rate": 0.0005504267730037283, + "loss": 4.213962554931641, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01688133366405964, + "learning_rate": 0.0005502792130337508, + "loss": 4.199461460113525, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.019166061654686928, + "learning_rate": 0.0005501314538353776, + "loss": 4.228936672210693, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.024065518751740456, + "learning_rate": 0.000549983495527656, + "loss": 4.214227676391602, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02518288791179657, + "learning_rate": 0.000549835338229793, + "loss": 4.256412029266357, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.026093967258930206, + "learning_rate": 0.0005496869820611563, + "loss": 4.216564178466797, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.023775635287165642, + "learning_rate": 0.0005495384271412739, + "loss": 4.226430892944336, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.018810907378792763, + "learning_rate": 0.0005493896735898338, + "loss": 4.190013885498047, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.018974872305989265, + "learning_rate": 0.000549240721526684, + "loss": 4.246231555938721, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.02069833129644394, + "learning_rate": 0.0005490915710718325, + "loss": 4.198457717895508, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.02097306028008461, + "learning_rate": 0.0005489422223454471, + "loss": 4.212507247924805, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022165542468428612, + "learning_rate": 0.0005487926754678556, + "loss": 4.2043843269348145, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.019068919122219086, + "learning_rate": 0.0005486429305595448, + "loss": 4.197460651397705, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01701812446117401, + "learning_rate": 0.0005484929877411618, + "loss": 4.220056533813477, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01629726029932499, + "learning_rate": 0.0005483428471335127, + "loss": 4.214868545532227, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018661806359887123, + "learning_rate": 0.000548192508857563, + "loss": 4.227047920227051, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.02005891129374504, + "learning_rate": 0.0005480419730344377, + "loss": 4.197265625, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.02591128647327423, + "learning_rate": 0.0005478912397854208, + "loss": 4.211041450500488, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.032659754157066345, + "learning_rate": 0.0005477403092319552, + "loss": 4.230195045471191, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.04057808965444565, + "learning_rate": 0.0005475891814956429, + "loss": 4.223843097686768, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.04103454574942589, + "learning_rate": 0.0005474378566982449, + "loss": 4.197145462036133, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.03810729831457138, + "learning_rate": 0.0005472863349616809, + "loss": 4.237957000732422, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.028838301077485085, + "learning_rate": 0.0005471346164080292, + "loss": 4.206722259521484, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02436191774904728, + "learning_rate": 0.0005469827011595266, + "loss": 4.1983137130737305, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.022839387878775597, + "learning_rate": 0.0005468305893385686, + "loss": 4.20530366897583, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.02375260926783085, + "learning_rate": 0.0005466782810677088, + "loss": 4.213202476501465, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.02177741937339306, + "learning_rate": 0.0005465257764696593, + "loss": 4.200539588928223, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.02351468615233898, + "learning_rate": 0.0005463730756672903, + "loss": 4.198091506958008, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.023375794291496277, + "learning_rate": 0.0005462201787836301, + "loss": 4.193244934082031, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.02479805052280426, + "learning_rate": 0.0005460670859418651, + "loss": 4.2250871658325195, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.026894841343164444, + "learning_rate": 0.000545913797265339, + "loss": 4.241791248321533, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.030582403764128685, + "learning_rate": 0.0005457603128775541, + "loss": 4.203666687011719, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.032222259789705276, + "learning_rate": 0.0005456066329021699, + "loss": 4.209629535675049, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.03181060031056404, + "learning_rate": 0.0005454527574630037, + "loss": 4.200459957122803, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02687576413154602, + "learning_rate": 0.0005452986866840298, + "loss": 4.211819648742676, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.025988668203353882, + "learning_rate": 0.0005451444206893806, + "loss": 4.198040962219238, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.02937682718038559, + "learning_rate": 0.0005449899596033453, + "loss": 4.222935676574707, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.03043951839208603, + "learning_rate": 0.0005448353035503704, + "loss": 4.2031755447387695, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.030401621013879776, + "learning_rate": 0.0005446804526550596, + "loss": 4.198380470275879, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 4.368444919586182, + "eval_runtime": 58.3009, + "eval_samples_per_second": 41.886, + "eval_steps_per_second": 1.321, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.03461585193872452, + "learning_rate": 0.0005445254070421732, + "loss": 4.147369384765625, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.039084456861019135, + "learning_rate": 0.0005443701668366288, + "loss": 4.124129295349121, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.04021848738193512, + "learning_rate": 0.0005442147321635007, + "loss": 4.127362251281738, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.041000768542289734, + "learning_rate": 0.00054405910314802, + "loss": 4.1432061195373535, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.04008613899350166, + "learning_rate": 0.0005439032799155737, + "loss": 4.169931888580322, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.039394307881593704, + "learning_rate": 0.000543747262591706, + "loss": 4.155813217163086, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.04374868795275688, + "learning_rate": 0.0005435910513021172, + "loss": 4.159891128540039, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0443582646548748, + "learning_rate": 0.0005434346461726641, + "loss": 4.157362937927246, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.04601646214723587, + "learning_rate": 0.0005432780473293593, + "loss": 4.152097225189209, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.04327242821455002, + "learning_rate": 0.0005431212548983718, + "loss": 4.170074462890625, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.0413462370634079, + "learning_rate": 0.0005429642690060262, + "loss": 4.156710147857666, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.036855220794677734, + "learning_rate": 0.0005428070897788034, + "loss": 4.138053894042969, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.03781217709183693, + "learning_rate": 0.0005426497173433398, + "loss": 4.172306060791016, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.03886301815509796, + "learning_rate": 0.0005424921518264275, + "loss": 4.135544776916504, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.03730994835495949, + "learning_rate": 0.0005423343933550143, + "loss": 4.171654224395752, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.03271950036287308, + "learning_rate": 0.0005421764420562032, + "loss": 4.169906139373779, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.03292673081159592, + "learning_rate": 0.0005420182980572527, + "loss": 4.163660049438477, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.029366906732320786, + "learning_rate": 0.0005418599614855768, + "loss": 4.161055088043213, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.023805437609553337, + "learning_rate": 0.0005417014324687442, + "loss": 4.172390937805176, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.023178137838840485, + "learning_rate": 0.000541542711134479, + "loss": 4.146788597106934, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.02403450198471546, + "learning_rate": 0.0005413837976106599, + "loss": 4.124364852905273, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.025009717792272568, + "learning_rate": 0.0005412246920253208, + "loss": 4.12275505065918, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.023431289941072464, + "learning_rate": 0.0005410653945066503, + "loss": 4.134739398956299, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.023803815245628357, + "learning_rate": 0.0005409059051829914, + "loss": 4.130463123321533, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.021219994872808456, + "learning_rate": 0.0005407462241828417, + "loss": 4.124577522277832, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.01899666152894497, + "learning_rate": 0.0005405863516348535, + "loss": 4.167086124420166, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.01889493502676487, + "learning_rate": 0.000540426287667833, + "loss": 4.152744293212891, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.019546914845705032, + "learning_rate": 0.000540266032410741, + "loss": 4.122655868530273, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02085448056459427, + "learning_rate": 0.0005401055859926923, + "loss": 4.141949653625488, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.019959403201937675, + "learning_rate": 0.0005399449485429555, + "loss": 4.131907939910889, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.01885392889380455, + "learning_rate": 0.0005397841201909535, + "loss": 4.118953227996826, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.018216131255030632, + "learning_rate": 0.0005396231010662627, + "loss": 4.096459865570068, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.019385207444429398, + "learning_rate": 0.0005394618912986132, + "loss": 4.111050128936768, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.018676603212952614, + "learning_rate": 0.0005393004910178889, + "loss": 4.156215667724609, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.01876339502632618, + "learning_rate": 0.0005391389003541271, + "loss": 4.1107282638549805, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.01831623539328575, + "learning_rate": 0.0005389771194375185, + "loss": 4.154116630554199, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.018896661698818207, + "learning_rate": 0.0005388151483984071, + "loss": 4.093320846557617, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.019582601264119148, + "learning_rate": 0.0005386529873672898, + "loss": 4.106500625610352, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.020313601940870285, + "learning_rate": 0.0005384906364748172, + "loss": 4.071961879730225, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.020385954529047012, + "learning_rate": 0.000538328095851792, + "loss": 4.129006862640381, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02231123298406601, + "learning_rate": 0.0005381653656291707, + "loss": 4.067756652832031, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.022957179695367813, + "learning_rate": 0.0005380024459380619, + "loss": 4.088230133056641, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021438462659716606, + "learning_rate": 0.000537839336909727, + "loss": 4.128036022186279, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01893841102719307, + "learning_rate": 0.00053767603867558, + "loss": 4.155407905578613, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.018539676442742348, + "learning_rate": 0.0005375125513671874, + "loss": 4.101880073547363, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.01881185919046402, + "learning_rate": 0.0005373488751162681, + "loss": 4.118906021118164, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.021657396107912064, + "learning_rate": 0.0005371850100546929, + "loss": 4.091935157775879, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.023293187841773033, + "learning_rate": 0.000537020956314485, + "loss": 4.105803489685059, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.021306317299604416, + "learning_rate": 0.0005368567140278196, + "loss": 4.0610809326171875, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.022078553214669228, + "learning_rate": 0.0005366922833270235, + "loss": 4.09867525100708, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.02266664244234562, + "learning_rate": 0.0005365276643445757, + "loss": 4.1027326583862305, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.02091757394373417, + "learning_rate": 0.0005363628572131069, + "loss": 4.080266952514648, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.020374659448862076, + "learning_rate": 0.0005361978620653991, + "loss": 4.088479042053223, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.02294750325381756, + "learning_rate": 0.0005360326790343858, + "loss": 4.084364891052246, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.02337353676557541, + "learning_rate": 0.000535867308253152, + "loss": 4.125696182250977, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.022643696516752243, + "learning_rate": 0.0005357017498549341, + "loss": 4.097160339355469, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.021080605685710907, + "learning_rate": 0.0005355360039731196, + "loss": 4.102363109588623, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.02269427664577961, + "learning_rate": 0.0005353700707412466, + "loss": 4.089844703674316, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.02195413038134575, + "learning_rate": 0.0005352039502930048, + "loss": 4.132293701171875, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.01805616170167923, + "learning_rate": 0.0005350376427622343, + "loss": 4.0842814445495605, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.018048783764243126, + "learning_rate": 0.0005348711482829261, + "loss": 4.109591007232666, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.0199508648365736, + "learning_rate": 0.000534704466989222, + "loss": 4.0908074378967285, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.020186785608530045, + "learning_rate": 0.0005345375990154138, + "loss": 4.077959060668945, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.022526193410158157, + "learning_rate": 0.0005343705444959441, + "loss": 4.109979629516602, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.025224441662430763, + "learning_rate": 0.000534203303565406, + "loss": 4.047533988952637, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.027292724698781967, + "learning_rate": 0.000534035876358542, + "loss": 4.09566593170166, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.025669090449810028, + "learning_rate": 0.0005338682630102454, + "loss": 4.0606160163879395, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02791150100529194, + "learning_rate": 0.0005337004636555595, + "loss": 4.097081184387207, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.03294133022427559, + "learning_rate": 0.0005335324784296768, + "loss": 4.121442794799805, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.027322066947817802, + "learning_rate": 0.0005333643074679403, + "loss": 4.1188740730285645, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.024866372346878052, + "learning_rate": 0.000533195950905842, + "loss": 4.082042694091797, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.02901027724146843, + "learning_rate": 0.0005330274088790239, + "loss": 4.110199451446533, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 4.283970355987549, + "eval_runtime": 58.4003, + "eval_samples_per_second": 41.815, + "eval_steps_per_second": 1.318, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.028958842158317566, + "learning_rate": 0.0005328586815232773, + "loss": 4.040733337402344, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.03491180017590523, + "learning_rate": 0.0005326897689745428, + "loss": 4.0473432540893555, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.04025818780064583, + "learning_rate": 0.0005325206713689101, + "loss": 4.021976470947266, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03998899459838867, + "learning_rate": 0.0005323513888426181, + "loss": 4.04901123046875, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.042165111750364304, + "learning_rate": 0.0005321819215320546, + "loss": 4.050230979919434, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.03511597216129303, + "learning_rate": 0.0005320122695737565, + "loss": 4.0989990234375, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.04064979776740074, + "learning_rate": 0.0005318424331044094, + "loss": 4.051448822021484, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.047536592930555344, + "learning_rate": 0.0005316724122608473, + "loss": 4.064278602600098, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.05592669919133186, + "learning_rate": 0.0005315022071800527, + "loss": 4.045516014099121, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.04262633994221687, + "learning_rate": 0.0005313318179991574, + "loss": 4.060666084289551, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03997781500220299, + "learning_rate": 0.0005311612448554402, + "loss": 4.065315246582031, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.038152433931827545, + "learning_rate": 0.0005309904878863293, + "loss": 4.069721221923828, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.0367623008787632, + "learning_rate": 0.0005308195472294001, + "loss": 4.080986976623535, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.0358598455786705, + "learning_rate": 0.0005306484230223769, + "loss": 4.065290927886963, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.039188165217638016, + "learning_rate": 0.000530477115403131, + "loss": 4.024333477020264, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.036839909851551056, + "learning_rate": 0.0005303056245096817, + "loss": 4.067532539367676, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.03456985577940941, + "learning_rate": 0.0005301339504801967, + "loss": 4.048443794250488, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.039487943053245544, + "learning_rate": 0.0005299620934529902, + "loss": 4.0585737228393555, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.03415827825665474, + "learning_rate": 0.0005297900535665244, + "loss": 4.06104850769043, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.03209620714187622, + "learning_rate": 0.0005296178309594087, + "loss": 4.009165287017822, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.028947360813617706, + "learning_rate": 0.0005294454257704, + "loss": 4.067384243011475, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026836825534701347, + "learning_rate": 0.0005292728381384018, + "loss": 4.046419620513916, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024095216765999794, + "learning_rate": 0.0005291000682024647, + "loss": 4.019008636474609, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.020965851843357086, + "learning_rate": 0.0005289271161017867, + "loss": 4.020613670349121, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.02250855043530464, + "learning_rate": 0.0005287539819757119, + "loss": 4.031634330749512, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.021385207772254944, + "learning_rate": 0.0005285806659637315, + "loss": 4.086246013641357, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.021104589104652405, + "learning_rate": 0.000528407168205483, + "loss": 4.033970355987549, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.019584037363529205, + "learning_rate": 0.0005282334888407503, + "loss": 4.025118827819824, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.02107258513569832, + "learning_rate": 0.000528059628009464, + "loss": 4.020823955535889, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.020438389852643013, + "learning_rate": 0.0005278855858517003, + "loss": 4.016323566436768, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.01881488971412182, + "learning_rate": 0.0005277113625076822, + "loss": 3.9991841316223145, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.01763618178665638, + "learning_rate": 0.000527536958117778, + "loss": 4.0062150955200195, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.017308862879872322, + "learning_rate": 0.0005273623728225022, + "loss": 3.946723461151123, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01706845872104168, + "learning_rate": 0.0005271876067625153, + "loss": 4.056562900543213, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.018664294853806496, + "learning_rate": 0.0005270126600786228, + "loss": 4.024652481079102, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.018523404374718666, + "learning_rate": 0.0005268375329117762, + "loss": 3.962064743041992, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.017408935353159904, + "learning_rate": 0.0005266622254030724, + "loss": 4.008697509765625, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.01588570512831211, + "learning_rate": 0.0005264867376937534, + "loss": 4.011664390563965, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.015084310434758663, + "learning_rate": 0.0005263110699252065, + "loss": 3.996224880218506, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01637054979801178, + "learning_rate": 0.000526135222238964, + "loss": 4.015972137451172, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.01673802360892296, + "learning_rate": 0.0005259591947767032, + "loss": 3.975414276123047, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.016544019803404808, + "learning_rate": 0.0005257829876802464, + "loss": 4.012955188751221, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.017455460503697395, + "learning_rate": 0.0005256066010915603, + "loss": 4.010180473327637, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.017625460401177406, + "learning_rate": 0.0005254300351527566, + "loss": 4.001679420471191, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.01732795126736164, + "learning_rate": 0.000525253290006091, + "loss": 4.006107330322266, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.016654646024107933, + "learning_rate": 0.0005250763657939641, + "loss": 4.045180797576904, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.015085658989846706, + "learning_rate": 0.0005248992626589203, + "loss": 3.994246482849121, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.015112568624317646, + "learning_rate": 0.0005247219807436489, + "loss": 4.004963397979736, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.014882124960422516, + "learning_rate": 0.000524544520190982, + "loss": 4.029565811157227, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.016012471169233322, + "learning_rate": 0.000524366881143897, + "loss": 4.018208026885986, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.015761924907565117, + "learning_rate": 0.0005241890637455141, + "loss": 3.9911205768585205, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.017770355567336082, + "learning_rate": 0.0005240110681390978, + "loss": 4.028733253479004, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01959751546382904, + "learning_rate": 0.0005238328944680558, + "loss": 4.003322124481201, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.018713824450969696, + "learning_rate": 0.0005236545428759394, + "loss": 3.991718292236328, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.01992208883166313, + "learning_rate": 0.0005234760135064434, + "loss": 4.002507209777832, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.021671291440725327, + "learning_rate": 0.0005232973065034055, + "loss": 4.021344184875488, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024282798171043396, + "learning_rate": 0.0005231184220108069, + "loss": 3.9899158477783203, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.023294344544410706, + "learning_rate": 0.0005229393601727713, + "loss": 4.046627044677734, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.02029707282781601, + "learning_rate": 0.000522760121133566, + "loss": 4.022454261779785, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.02151239849627018, + "learning_rate": 0.0005225807050376004, + "loss": 4.005476951599121, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.02320881187915802, + "learning_rate": 0.0005224011120294269, + "loss": 4.012726783752441, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.026484191417694092, + "learning_rate": 0.0005222213422537401, + "loss": 3.9904391765594482, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.026279661804437637, + "learning_rate": 0.0005220413958553777, + "loss": 3.9738848209381104, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.024299729615449905, + "learning_rate": 0.0005218612729793188, + "loss": 4.003316402435303, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.02337595447897911, + "learning_rate": 0.0005216809737706856, + "loss": 3.954984188079834, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.02416428178548813, + "learning_rate": 0.0005215004983747415, + "loss": 4.0317063331604, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.022488107904791832, + "learning_rate": 0.0005213198469368925, + "loss": 3.9736647605895996, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.019700225442647934, + "learning_rate": 0.0005211390196026861, + "loss": 3.979867458343506, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.020382333546876907, + "learning_rate": 0.0005209580165178117, + "loss": 3.995251178741455, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.020149298012256622, + "learning_rate": 0.0005207768378281003, + "loss": 3.9987268447875977, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.019790509715676308, + "learning_rate": 0.0005205954836795241, + "loss": 3.9838991165161133, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.022096451371908188, + "learning_rate": 0.000520413954218197, + "loss": 3.995800256729126, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.214347839355469, + "eval_runtime": 58.8872, + "eval_samples_per_second": 41.469, + "eval_steps_per_second": 1.308, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02203402668237686, + "learning_rate": 0.0005202322495903739, + "loss": 3.915332078933716, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.02336675301194191, + "learning_rate": 0.0005200503699424509, + "loss": 3.914801597595215, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.027339637279510498, + "learning_rate": 0.0005198683154209654, + "loss": 3.9277780055999756, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.026872653514146805, + "learning_rate": 0.0005196860861725952, + "loss": 3.9160094261169434, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.026558255776762962, + "learning_rate": 0.0005195036823441592, + "loss": 3.9084300994873047, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.02644367143511772, + "learning_rate": 0.0005193211040826169, + "loss": 3.9328718185424805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.024677276611328125, + "learning_rate": 0.0005191383515350681, + "loss": 3.88993501663208, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.023194244131445885, + "learning_rate": 0.0005189554248487536, + "loss": 3.923567056655884, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.024653147906064987, + "learning_rate": 0.0005187723241710539, + "loss": 3.8740999698638916, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.026041202247142792, + "learning_rate": 0.0005185890496494899, + "loss": 3.9542148113250732, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.029586758464574814, + "learning_rate": 0.0005184056014317225, + "loss": 3.9241347312927246, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.0325787179172039, + "learning_rate": 0.0005182219796655528, + "loss": 3.9656593799591064, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.03183380514383316, + "learning_rate": 0.0005180381844989214, + "loss": 3.950648307800293, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.03434471786022186, + "learning_rate": 0.0005178542160799088, + "loss": 3.9590439796447754, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.030099432915449142, + "learning_rate": 0.0005176700745567353, + "loss": 3.9518113136291504, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.025950564071536064, + "learning_rate": 0.0005174857600777599, + "loss": 3.8908939361572266, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.029959935694932938, + "learning_rate": 0.0005173012727914816, + "loss": 3.945037364959717, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.029551763087511063, + "learning_rate": 0.0005171166128465389, + "loss": 3.923959732055664, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.03538523241877556, + "learning_rate": 0.0005169317803917083, + "loss": 3.9395527839660645, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.037924155592918396, + "learning_rate": 0.0005167467755759065, + "loss": 3.945115089416504, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.036773014813661575, + "learning_rate": 0.0005165615985481881, + "loss": 3.9484646320343018, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.04125838354229927, + "learning_rate": 0.0005163762494577474, + "loss": 4.000699043273926, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.044125135987997055, + "learning_rate": 0.0005161907284539165, + "loss": 3.9828028678894043, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.040853384882211685, + "learning_rate": 0.0005160050356861663, + "loss": 3.957505702972412, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.03501008450984955, + "learning_rate": 0.0005158191713041063, + "loss": 3.9360849857330322, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.04082556813955307, + "learning_rate": 0.0005156331354574838, + "loss": 3.95941424369812, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.045432962477207184, + "learning_rate": 0.0005154469282961847, + "loss": 3.9666922092437744, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.04048244655132294, + "learning_rate": 0.0005152605499702328, + "loss": 3.984555244445801, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.031016314402222633, + "learning_rate": 0.0005150740006297896, + "loss": 3.945204257965088, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.03210771828889847, + "learning_rate": 0.0005148872804251548, + "loss": 3.976186752319336, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.03130023553967476, + "learning_rate": 0.000514700389506765, + "loss": 3.915727138519287, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.031193193048238754, + "learning_rate": 0.0005145133280251954, + "loss": 3.961866855621338, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.027518954128026962, + "learning_rate": 0.0005143260961311578, + "loss": 3.9742956161499023, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.02751155197620392, + "learning_rate": 0.0005141386939755014, + "loss": 3.949662208557129, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.026033353060483932, + "learning_rate": 0.0005139511217092129, + "loss": 3.9327926635742188, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.02579621411859989, + "learning_rate": 0.000513763379483416, + "loss": 3.9334418773651123, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.023009559139609337, + "learning_rate": 0.0005135754674493709, + "loss": 3.969707489013672, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.02196447364985943, + "learning_rate": 0.0005133873857584753, + "loss": 3.9084434509277344, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.02206484042108059, + "learning_rate": 0.0005131991345622629, + "loss": 3.922117233276367, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.023360205814242363, + "learning_rate": 0.0005130107140124046, + "loss": 3.965158224105835, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02162669226527214, + "learning_rate": 0.0005128221242607072, + "loss": 3.9519729614257812, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.019719846546649933, + "learning_rate": 0.0005126333654591141, + "loss": 3.9195961952209473, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.018988260999321938, + "learning_rate": 0.0005124444377597049, + "loss": 3.937926769256592, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.019278401508927345, + "learning_rate": 0.0005122553413146951, + "loss": 3.9126808643341064, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019101005047559738, + "learning_rate": 0.0005120660762764366, + "loss": 3.8993749618530273, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.019156700000166893, + "learning_rate": 0.0005118766427974166, + "loss": 3.9531068801879883, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019944407045841217, + "learning_rate": 0.0005116870410302582, + "loss": 3.9645180702209473, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.01826174184679985, + "learning_rate": 0.0005114972711277202, + "loss": 3.948477268218994, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.0191368218511343, + "learning_rate": 0.0005113073332426968, + "loss": 3.944648027420044, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.017768343910574913, + "learning_rate": 0.0005111172275282174, + "loss": 3.9584474563598633, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.01955590397119522, + "learning_rate": 0.0005109269541374469, + "loss": 3.8936800956726074, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.02100536972284317, + "learning_rate": 0.000510736513223685, + "loss": 3.93919038772583, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.020873935893177986, + "learning_rate": 0.0005105459049403665, + "loss": 3.929244041442871, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.021876046434044838, + "learning_rate": 0.0005103551294410614, + "loss": 3.9264731407165527, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.021726684644818306, + "learning_rate": 0.0005101641868794736, + "loss": 3.923135280609131, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.019308188930153847, + "learning_rate": 0.0005099730774094423, + "loss": 3.9062135219573975, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01917017623782158, + "learning_rate": 0.0005097818011849408, + "loss": 3.898810386657715, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017933931201696396, + "learning_rate": 0.0005095903583600772, + "loss": 3.8899734020233154, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.015131795778870583, + "learning_rate": 0.0005093987490890933, + "loss": 3.9452900886535645, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.017190825194120407, + "learning_rate": 0.0005092069735263651, + "loss": 3.9124746322631836, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.01618487387895584, + "learning_rate": 0.000509015031826403, + "loss": 3.91654109954834, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.014920198358595371, + "learning_rate": 0.0005088229241438509, + "loss": 3.9144535064697266, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01615205965936184, + "learning_rate": 0.0005086306506334864, + "loss": 3.917003631591797, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.015370192006230354, + "learning_rate": 0.0005084382114502208, + "loss": 3.8910603523254395, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.015290996059775352, + "learning_rate": 0.0005082456067490988, + "loss": 3.9122695922851562, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.015913106501102448, + "learning_rate": 0.0005080528366852987, + "loss": 3.9304556846618652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.01639600843191147, + "learning_rate": 0.0005078599014141318, + "loss": 3.8974502086639404, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.017010601237416267, + "learning_rate": 0.0005076668010910425, + "loss": 3.9723148345947266, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.017778996378183365, + "learning_rate": 0.0005074735358716083, + "loss": 3.8905391693115234, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.020029611885547638, + "learning_rate": 0.0005072801059115394, + "loss": 3.934976100921631, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.01931857503950596, + "learning_rate": 0.000507086511366679, + "loss": 3.9152774810791016, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019750485196709633, + "learning_rate": 0.0005068927523930027, + "loss": 3.877650499343872, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.161914825439453, + "eval_runtime": 58.5217, + "eval_samples_per_second": 41.728, + "eval_steps_per_second": 1.316, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02073049545288086, + "learning_rate": 0.0005066988291466184, + "loss": 3.8331246376037598, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02272389829158783, + "learning_rate": 0.0005065047417837669, + "loss": 3.8847155570983887, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02139078453183174, + "learning_rate": 0.0005063104904608206, + "loss": 3.8276169300079346, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02043229155242443, + "learning_rate": 0.0005061160753342843, + "loss": 3.8804001808166504, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.02345564216375351, + "learning_rate": 0.0005059214965607948, + "loss": 3.851609468460083, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.023448185995221138, + "learning_rate": 0.0005057267542971207, + "loss": 3.8446507453918457, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.024826066568493843, + "learning_rate": 0.0005055318487001624, + "loss": 3.86712384223938, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.02491980977356434, + "learning_rate": 0.0005053367799269518, + "loss": 3.8374528884887695, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.023002834990620613, + "learning_rate": 0.0005051415481346522, + "loss": 3.783214569091797, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.02123952843248844, + "learning_rate": 0.0005049461534805585, + "loss": 3.8463428020477295, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.01981225423514843, + "learning_rate": 0.0005047505961220964, + "loss": 3.8307082653045654, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01915367692708969, + "learning_rate": 0.0005045548762168231, + "loss": 3.8190226554870605, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.01890859752893448, + "learning_rate": 0.0005043589939224266, + "loss": 3.8428220748901367, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.019150376319885254, + "learning_rate": 0.0005041629493967257, + "loss": 3.830247402191162, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.020099453628063202, + "learning_rate": 0.0005039667427976697, + "loss": 3.838806629180908, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.021136781200766563, + "learning_rate": 0.0005037703742833391, + "loss": 3.8344273567199707, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.022857150062918663, + "learning_rate": 0.0005035738440119442, + "loss": 3.8317880630493164, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.024891739711165428, + "learning_rate": 0.0005033771521418261, + "loss": 3.837855577468872, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.02595464698970318, + "learning_rate": 0.0005031802988314557, + "loss": 3.846050262451172, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.02722409926354885, + "learning_rate": 0.0005029832842394342, + "loss": 3.8164305686950684, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.024486860260367393, + "learning_rate": 0.0005027861085244927, + "loss": 3.8099875450134277, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.021551169455051422, + "learning_rate": 0.0005025887718454922, + "loss": 3.825925827026367, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.02121506817638874, + "learning_rate": 0.0005023912743614232, + "loss": 3.870727777481079, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.01948230341076851, + "learning_rate": 0.0005021936162314058, + "loss": 3.8035314083099365, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.01896044984459877, + "learning_rate": 0.0005019957976146898, + "loss": 3.84665584564209, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.020620400086045265, + "learning_rate": 0.0005017978186706538, + "loss": 3.8338310718536377, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.020855262875556946, + "learning_rate": 0.0005015996795588062, + "loss": 3.8781003952026367, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.02153870463371277, + "learning_rate": 0.0005014013804387838, + "loss": 3.8783490657806396, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.023751024156808853, + "learning_rate": 0.0005012029214703529, + "loss": 3.846198558807373, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.023987066000699997, + "learning_rate": 0.000501004302813408, + "loss": 3.834296226501465, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.023626182228326797, + "learning_rate": 0.0005008055246279727, + "loss": 3.8591227531433105, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.025707192718982697, + "learning_rate": 0.000500606587074199, + "loss": 3.8766489028930664, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.027723029255867004, + "learning_rate": 0.0005004074903123674, + "loss": 3.88264799118042, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.0278895553201437, + "learning_rate": 0.0005002082345028864, + "loss": 3.8932995796203613, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.025622662156820297, + "learning_rate": 0.0005000088198062929, + "loss": 3.8734612464904785, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.02639615908265114, + "learning_rate": 0.0004998092463832516, + "loss": 3.786813735961914, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.026765599846839905, + "learning_rate": 0.0004996095143945554, + "loss": 3.853966474533081, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02823319099843502, + "learning_rate": 0.0004994096240011246, + "loss": 3.8321340084075928, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.023394955322146416, + "learning_rate": 0.0004992095753640072, + "loss": 3.8527331352233887, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.023874662816524506, + "learning_rate": 0.0004990093686443791, + "loss": 3.8822879791259766, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.02301311306655407, + "learning_rate": 0.000498809004003543, + "loss": 3.846186876296997, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.022630592808127403, + "learning_rate": 0.0004986084816029291, + "loss": 3.867992639541626, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.02424517087638378, + "learning_rate": 0.0004984078016040946, + "loss": 3.867293357849121, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.024116624146699905, + "learning_rate": 0.000498206964168724, + "loss": 3.846282958984375, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.02187768742442131, + "learning_rate": 0.000498005969458628, + "loss": 3.8429484367370605, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.01879809983074665, + "learning_rate": 0.0004978048176357447, + "loss": 3.863215446472168, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.020426234230399132, + "learning_rate": 0.0004976035088621385, + "loss": 3.825230121612549, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01992812193930149, + "learning_rate": 0.0004974020433000003, + "loss": 3.851665496826172, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.020693328231573105, + "learning_rate": 0.000497200421111647, + "loss": 3.906999111175537, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.021137919276952744, + "learning_rate": 0.0004969986424595221, + "loss": 3.8369035720825195, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.023285407572984695, + "learning_rate": 0.0004967967075061949, + "loss": 3.8658390045166016, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.0219135619699955, + "learning_rate": 0.0004965946164143609, + "loss": 3.873072862625122, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.020101983100175858, + "learning_rate": 0.0004963923693468413, + "loss": 3.847805976867676, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.01875176839530468, + "learning_rate": 0.0004961899664665827, + "loss": 3.7976622581481934, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01913360133767128, + "learning_rate": 0.0004959874079366576, + "loss": 3.853851318359375, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.016863619908690453, + "learning_rate": 0.0004957846939202637, + "loss": 3.877378225326538, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.017802366986870766, + "learning_rate": 0.000495581824580724, + "loss": 3.861356019973755, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.01981814205646515, + "learning_rate": 0.0004953788000814866, + "loss": 3.88613224029541, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.017508897930383682, + "learning_rate": 0.000495175620586125, + "loss": 3.8202030658721924, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.019240548834204674, + "learning_rate": 0.0004949722862583368, + "loss": 3.853827714920044, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.022305598482489586, + "learning_rate": 0.000494768797261945, + "loss": 3.885148763656616, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.02531025931239128, + "learning_rate": 0.000494565153760897, + "loss": 3.8412041664123535, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.024474849924445152, + "learning_rate": 0.0004943613559192648, + "loss": 3.865644931793213, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.020430924370884895, + "learning_rate": 0.0004941574039012444, + "loss": 3.8649442195892334, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.01906431093811989, + "learning_rate": 0.0004939532978711566, + "loss": 3.8462204933166504, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.021118057891726494, + "learning_rate": 0.0004937490379934456, + "loss": 3.8652522563934326, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.020337693393230438, + "learning_rate": 0.00049354462443268, + "loss": 3.8506221771240234, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.019090797752141953, + "learning_rate": 0.0004933400573535521, + "loss": 3.881237506866455, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.020973797887563705, + "learning_rate": 0.000493135336920878, + "loss": 3.821274757385254, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.021984362974762917, + "learning_rate": 0.0004929304632995973, + "loss": 3.8793835639953613, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.0215684212744236, + "learning_rate": 0.0004927254366547727, + "loss": 3.8259663581848145, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.021673865616321564, + "learning_rate": 0.0004925202571515908, + "loss": 3.8505711555480957, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.123871326446533, + "eval_runtime": 58.8357, + "eval_samples_per_second": 41.505, + "eval_steps_per_second": 1.309, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.0203359667211771, + "learning_rate": 0.0004923149249553608, + "loss": 3.785353660583496, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.023197868838906288, + "learning_rate": 0.0004921094402315152, + "loss": 3.7479515075683594, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.02314298041164875, + "learning_rate": 0.0004919038031456093, + "loss": 3.7697315216064453, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.0256908368319273, + "learning_rate": 0.0004916980138633213, + "loss": 3.79056978225708, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.026193121448159218, + "learning_rate": 0.0004914920725504518, + "loss": 3.7722532749176025, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.027120303362607956, + "learning_rate": 0.000491285979372924, + "loss": 3.7991981506347656, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.0266975536942482, + "learning_rate": 0.0004910797344967834, + "loss": 3.778301477432251, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.02806396596133709, + "learning_rate": 0.000490873338088198, + "loss": 3.7578704357147217, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.02893839031457901, + "learning_rate": 0.0004906667903134573, + "loss": 3.7401275634765625, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.031170370057225227, + "learning_rate": 0.0004904600913389735, + "loss": 3.780686616897583, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.03364266827702522, + "learning_rate": 0.0004902532413312799, + "loss": 3.794442653656006, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.030121006071567535, + "learning_rate": 0.0004900462404570321, + "loss": 3.779569149017334, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.03296204283833504, + "learning_rate": 0.0004898390888830067, + "loss": 3.7960336208343506, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.03381749242544174, + "learning_rate": 0.0004896317867761022, + "loss": 3.7959280014038086, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.035383883863687515, + "learning_rate": 0.000489424334303338, + "loss": 3.8008625507354736, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.03759761154651642, + "learning_rate": 0.0004892167316318548, + "loss": 3.7997028827667236, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.030651796609163284, + "learning_rate": 0.0004890089789289147, + "loss": 3.81191349029541, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.0262477844953537, + "learning_rate": 0.0004888010763618997, + "loss": 3.8046622276306152, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.030398324131965637, + "learning_rate": 0.0004885930240983136, + "loss": 3.7903637886047363, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.030515259131789207, + "learning_rate": 0.0004883848223057801, + "loss": 3.8111231327056885, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.031686048954725266, + "learning_rate": 0.00048817647115204386, + "loss": 3.781445026397705, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.025968654081225395, + "learning_rate": 0.00048796797080496936, + "loss": 3.788196563720703, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.025340190157294273, + "learning_rate": 0.00048775932143254176, + "loss": 3.763331413269043, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.024738596752285957, + "learning_rate": 0.0004875505232028661, + "loss": 3.8005969524383545, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.021990075707435608, + "learning_rate": 0.0004873415762841672, + "loss": 3.795844078063965, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.02191784791648388, + "learning_rate": 0.00048713248084478986, + "loss": 3.8062384128570557, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.023678259924054146, + "learning_rate": 0.0004869232370531986, + "loss": 3.7760679721832275, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.024137988686561584, + "learning_rate": 0.0004867138450779773, + "loss": 3.7917492389678955, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.02673223614692688, + "learning_rate": 0.0004865043050878293, + "loss": 3.8006649017333984, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.02395169623196125, + "learning_rate": 0.0004862946172515771, + "loss": 3.7902212142944336, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.02410844899713993, + "learning_rate": 0.0004860847817381625, + "loss": 3.8073291778564453, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.028109831735491753, + "learning_rate": 0.0004858747987166463, + "loss": 3.80116605758667, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.025538964197039604, + "learning_rate": 0.00048566466835620774, + "loss": 3.791151523590088, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.02393815852701664, + "learning_rate": 0.0004854543908261453, + "loss": 3.7766106128692627, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.024659741669893265, + "learning_rate": 0.0004852439662958756, + "loss": 3.789135217666626, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.02668776921927929, + "learning_rate": 0.000485033394934934, + "loss": 3.797905445098877, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.0258641317486763, + "learning_rate": 0.000484822676912974, + "loss": 3.783323287963867, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.024846022948622704, + "learning_rate": 0.0004846118123997674, + "loss": 3.7976717948913574, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.024981874972581863, + "learning_rate": 0.00048440080156520387, + "loss": 3.818732738494873, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.02556881681084633, + "learning_rate": 0.0004841896445792909, + "loss": 3.7821896076202393, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.02558571845293045, + "learning_rate": 0.000483978341612154, + "loss": 3.8143680095672607, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.02361566200852394, + "learning_rate": 0.0004837668928340362, + "loss": 3.7710189819335938, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022570716217160225, + "learning_rate": 0.0004835552984152978, + "loss": 3.775564670562744, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.022478004917502403, + "learning_rate": 0.00048334355852641664, + "loss": 3.7920689582824707, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021241728216409683, + "learning_rate": 0.0004831316733379879, + "loss": 3.7857837677001953, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.018921511247754097, + "learning_rate": 0.0004829196430207235, + "loss": 3.7659268379211426, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01926037296652794, + "learning_rate": 0.00048270746774545237, + "loss": 3.7527129650115967, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.018743667751550674, + "learning_rate": 0.0004824951476831204, + "loss": 3.767991065979004, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.017273949459195137, + "learning_rate": 0.00048228268300479015, + "loss": 3.7834174633026123, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.017054244875907898, + "learning_rate": 0.0004820700738816403, + "loss": 3.7775752544403076, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017190931364893913, + "learning_rate": 0.0004818573204849664, + "loss": 3.7709648609161377, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01665034517645836, + "learning_rate": 0.00048164442298617987, + "loss": 3.7575321197509766, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.016932843253016472, + "learning_rate": 0.00048143138155680844, + "loss": 3.7922799587249756, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01702750287950039, + "learning_rate": 0.0004812181963684958, + "loss": 3.7729780673980713, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01766599342226982, + "learning_rate": 0.00048100486759300134, + "loss": 3.7876572608947754, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01785697042942047, + "learning_rate": 0.00048079139540220024, + "loss": 3.8131699562072754, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01830688677728176, + "learning_rate": 0.0004805777799680831, + "loss": 3.790778636932373, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017348438501358032, + "learning_rate": 0.0004803640214627561, + "loss": 3.7984330654144287, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017999377101659775, + "learning_rate": 0.0004801501200584406, + "loss": 3.7756147384643555, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.01801210641860962, + "learning_rate": 0.000479936075927473, + "loss": 3.789759874343872, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.018445398658514023, + "learning_rate": 0.000479721889242305, + "loss": 3.7683000564575195, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.017544370144605637, + "learning_rate": 0.0004795075601755027, + "loss": 3.7776038646698, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.019834233447909355, + "learning_rate": 0.00047929308889974737, + "loss": 3.7559571266174316, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01791626401245594, + "learning_rate": 0.0004790784755878346, + "loss": 3.7746729850769043, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.017638562247157097, + "learning_rate": 0.00047886372041267434, + "loss": 3.7933082580566406, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.018974050879478455, + "learning_rate": 0.0004786488235472909, + "loss": 3.795713424682617, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.018694303929805756, + "learning_rate": 0.0004784337851648231, + "loss": 3.7675955295562744, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.01788092777132988, + "learning_rate": 0.0004782186054385233, + "loss": 3.786083698272705, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.01902497000992298, + "learning_rate": 0.0004780032845417579, + "loss": 3.7976233959198, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.020558880642056465, + "learning_rate": 0.000477787822648007, + "loss": 3.7500152587890625, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.020093899220228195, + "learning_rate": 0.00047757221993086443, + "loss": 3.703448534011841, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.019833799451589584, + "learning_rate": 0.00047735647656403754, + "loss": 3.756059408187866, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.093987941741943, + "eval_runtime": 57.6894, + "eval_samples_per_second": 42.33, + "eval_steps_per_second": 1.335, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.022167526185512543, + "learning_rate": 0.0004771405927213467, + "loss": 3.696444511413574, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.027874166145920753, + "learning_rate": 0.0004769245685767255, + "loss": 3.713630199432373, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.030872784554958344, + "learning_rate": 0.00047670840430422113, + "loss": 3.678295612335205, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.03246859833598137, + "learning_rate": 0.0004764921000779928, + "loss": 3.7186388969421387, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.03667069226503372, + "learning_rate": 0.0004762756560723131, + "loss": 3.7457995414733887, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.036286015063524246, + "learning_rate": 0.0004760590724615671, + "loss": 3.6881775856018066, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.03413614630699158, + "learning_rate": 0.0004758423494202522, + "loss": 3.6999263763427734, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.03384552150964737, + "learning_rate": 0.0004756254871229782, + "loss": 3.6959803104400635, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.03227181360125542, + "learning_rate": 0.00047540848574446724, + "loss": 3.7182722091674805, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.036366235464811325, + "learning_rate": 0.0004751913454595535, + "loss": 3.7116851806640625, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.03573326766490936, + "learning_rate": 0.0004749740664431828, + "loss": 3.741865873336792, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.03482096269726753, + "learning_rate": 0.000474756648870413, + "loss": 3.7195520401000977, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.03232625126838684, + "learning_rate": 0.0004745390929164134, + "loss": 3.7443251609802246, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.03421664610505104, + "learning_rate": 0.0004743213987564651, + "loss": 3.6867361068725586, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.03585013002157211, + "learning_rate": 0.0004741035665659603, + "loss": 3.7492403984069824, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.029985329136252403, + "learning_rate": 0.00047388559652040236, + "loss": 3.705209493637085, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.02729589305818081, + "learning_rate": 0.0004736674887954059, + "loss": 3.711106538772583, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.029642216861248016, + "learning_rate": 0.0004734492435666963, + "loss": 3.7274727821350098, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.02816903032362461, + "learning_rate": 0.0004732308610101099, + "loss": 3.7181742191314697, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.027143625542521477, + "learning_rate": 0.00047301234130159346, + "loss": 3.719235897064209, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.025802744552493095, + "learning_rate": 0.00047279368461720445, + "loss": 3.771820545196533, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.026424532756209373, + "learning_rate": 0.0004725748911331106, + "loss": 3.7516651153564453, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.029677648097276688, + "learning_rate": 0.0004723559610255898, + "loss": 3.7058472633361816, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.031548649072647095, + "learning_rate": 0.0004721368944710302, + "loss": 3.756992816925049, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.027667395770549774, + "learning_rate": 0.00047191769164592974, + "loss": 3.742711067199707, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.027639545500278473, + "learning_rate": 0.000471698352726896, + "loss": 3.728883743286133, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.02950332500040531, + "learning_rate": 0.0004714788778906467, + "loss": 3.7160212993621826, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.026302607730031013, + "learning_rate": 0.0004712592673140084, + "loss": 3.735713481903076, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.022483112290501595, + "learning_rate": 0.00047103952117391764, + "loss": 3.702686309814453, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.021729422733187675, + "learning_rate": 0.00047081963964741986, + "loss": 3.7195816040039062, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.021159254014492035, + "learning_rate": 0.0004705996229116696, + "loss": 3.684321880340576, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.02233424223959446, + "learning_rate": 0.0004703794711439304, + "loss": 3.7233057022094727, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.02142048068344593, + "learning_rate": 0.0004701591845215744, + "loss": 3.7083680629730225, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.022331461310386658, + "learning_rate": 0.00046993876322208274, + "loss": 3.7408461570739746, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.020629866048693657, + "learning_rate": 0.0004697182074230448, + "loss": 3.759129047393799, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.02174685336649418, + "learning_rate": 0.0004694975173021586, + "loss": 3.72351336479187, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.021538110449910164, + "learning_rate": 0.0004692766930372299, + "loss": 3.716416597366333, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.020800625905394554, + "learning_rate": 0.00046905573480617285, + "loss": 3.726954936981201, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.02039588987827301, + "learning_rate": 0.00046883464278700966, + "loss": 3.7336390018463135, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.02107219770550728, + "learning_rate": 0.0004686134171578702, + "loss": 3.722311496734619, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.02068956382572651, + "learning_rate": 0.00046839205809699183, + "loss": 3.767139434814453, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.020190054550766945, + "learning_rate": 0.0004681705657827195, + "loss": 3.6960511207580566, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.018814411014318466, + "learning_rate": 0.0004679489403935058, + "loss": 3.70375919342041, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.017568519338965416, + "learning_rate": 0.00046772718210791005, + "loss": 3.756627321243286, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.018188195303082466, + "learning_rate": 0.00046750529110459903, + "loss": 3.6964683532714844, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.018925746902823448, + "learning_rate": 0.0004672832675623463, + "loss": 3.731172561645508, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.01690756157040596, + "learning_rate": 0.0004670611116600322, + "loss": 3.6955318450927734, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016300756484270096, + "learning_rate": 0.00046683882357664375, + "loss": 3.762632369995117, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.016405068337917328, + "learning_rate": 0.0004666164034912744, + "loss": 3.75130558013916, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.017687877640128136, + "learning_rate": 0.00046639385158312406, + "loss": 3.687530755996704, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.017524344846606255, + "learning_rate": 0.00046617116803149874, + "loss": 3.738790273666382, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.01714804768562317, + "learning_rate": 0.00046594835301581053, + "loss": 3.69827938079834, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018042324110865593, + "learning_rate": 0.00046572540671557745, + "loss": 3.71600341796875, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018648479133844376, + "learning_rate": 0.00046550232931042346, + "loss": 3.742591619491577, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.019198937341570854, + "learning_rate": 0.00046527912098007787, + "loss": 3.739820957183838, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.01796373911201954, + "learning_rate": 0.00046505578190437564, + "loss": 3.719025135040283, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01836620457470417, + "learning_rate": 0.0004648323122632569, + "loss": 3.741434097290039, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.02192627638578415, + "learning_rate": 0.0004646087122367673, + "loss": 3.7553882598876953, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.02147185243666172, + "learning_rate": 0.00046438498200505727, + "loss": 3.7705790996551514, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01855960488319397, + "learning_rate": 0.0004641611217483821, + "loss": 3.704925537109375, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.01925475336611271, + "learning_rate": 0.00046393713164710217, + "loss": 3.711491346359253, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019796503707766533, + "learning_rate": 0.00046371301188168204, + "loss": 3.7241992950439453, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.018511971458792686, + "learning_rate": 0.0004634887626326911, + "loss": 3.6785435676574707, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.018214019015431404, + "learning_rate": 0.00046326438408080293, + "loss": 3.7299962043762207, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.018912190571427345, + "learning_rate": 0.00046303987640679517, + "loss": 3.773698329925537, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.01939854398369789, + "learning_rate": 0.0004628152397915498, + "loss": 3.682321071624756, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.017859861254692078, + "learning_rate": 0.00046259047441605215, + "loss": 3.7254951000213623, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.019249707460403442, + "learning_rate": 0.0004623655804613919, + "loss": 3.7263433933258057, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.019221967086195946, + "learning_rate": 0.00046214055810876194, + "loss": 3.7229225635528564, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01871953159570694, + "learning_rate": 0.00046191540753945886, + "loss": 3.7248311042785645, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01747182011604309, + "learning_rate": 0.00046169012893488204, + "loss": 3.699913740158081, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.019630730152130127, + "learning_rate": 0.00046146472247653484, + "loss": 3.787783145904541, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.072091102600098, + "eval_runtime": 63.1589, + "eval_samples_per_second": 38.664, + "eval_steps_per_second": 1.219, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.022918934002518654, + "learning_rate": 0.0004612391883460228, + "loss": 3.666675567626953, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.0264838095754385, + "learning_rate": 0.00046101352672505493, + "loss": 3.6371259689331055, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.028463594615459442, + "learning_rate": 0.0004607877377954426, + "loss": 3.657130241394043, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.02900792844593525, + "learning_rate": 0.00046056182173909995, + "loss": 3.6365628242492676, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.028593607246875763, + "learning_rate": 0.0004603357787380435, + "loss": 3.647258996963501, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.02878790907561779, + "learning_rate": 0.0004601096089743919, + "loss": 3.682035446166992, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.02980183996260166, + "learning_rate": 0.0004598833126303661, + "loss": 3.623297691345215, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.03375725448131561, + "learning_rate": 0.00045965688988828884, + "loss": 3.6240830421447754, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.035045325756073, + "learning_rate": 0.00045943034093058507, + "loss": 3.6372249126434326, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.03519923985004425, + "learning_rate": 0.000459203665939781, + "loss": 3.6157641410827637, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.03468473628163338, + "learning_rate": 0.00045897686509850487, + "loss": 3.6553335189819336, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.034817393869161606, + "learning_rate": 0.0004587499385894856, + "loss": 3.653806686401367, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.03549059107899666, + "learning_rate": 0.000458522886595554, + "loss": 3.6730189323425293, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.028001410886645317, + "learning_rate": 0.0004582957092996418, + "loss": 3.662490129470825, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.02532520331442356, + "learning_rate": 0.0004580684068847817, + "loss": 3.629936456680298, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.028215918689966202, + "learning_rate": 0.0004578409795341069, + "loss": 3.6507444381713867, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.026823483407497406, + "learning_rate": 0.0004576134274308517, + "loss": 3.661811113357544, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.0281058382242918, + "learning_rate": 0.0004573857507583506, + "loss": 3.6805872917175293, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.02548503689467907, + "learning_rate": 0.00045715794970003874, + "loss": 3.6586413383483887, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.024795135483145714, + "learning_rate": 0.0004569300244394513, + "loss": 3.654989004135132, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.025501688942313194, + "learning_rate": 0.0004567019751602234, + "loss": 3.6915903091430664, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.02377370186150074, + "learning_rate": 0.00045647380204609034, + "loss": 3.660588264465332, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.022150425240397453, + "learning_rate": 0.00045624550528088704, + "loss": 3.64870023727417, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.02415974996984005, + "learning_rate": 0.00045601708504854803, + "loss": 3.6498191356658936, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.02140078693628311, + "learning_rate": 0.0004557885415331074, + "loss": 3.703489065170288, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.022121304646134377, + "learning_rate": 0.00045555987491869865, + "loss": 3.6606979370117188, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.02168990671634674, + "learning_rate": 0.00045533108538955416, + "loss": 3.6801178455352783, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.020228713750839233, + "learning_rate": 0.0004551021731300055, + "loss": 3.6707205772399902, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.018500352278351784, + "learning_rate": 0.0004548731383244834, + "loss": 3.647545576095581, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.020223816856741905, + "learning_rate": 0.0004546439811575169, + "loss": 3.644932985305786, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.019502121955156326, + "learning_rate": 0.00045441470181373375, + "loss": 3.62888240814209, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.01926770992577076, + "learning_rate": 0.0004541853004778603, + "loss": 3.675457715988159, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01888471283018589, + "learning_rate": 0.00045395577733472087, + "loss": 3.636434316635132, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.0201494712382555, + "learning_rate": 0.0004537261325692383, + "loss": 3.679147720336914, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.02178201824426651, + "learning_rate": 0.0004534963663664332, + "loss": 3.671905040740967, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.024073559790849686, + "learning_rate": 0.00045326647891142394, + "loss": 3.6701040267944336, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.02135559171438217, + "learning_rate": 0.0004530364703894268, + "loss": 3.648519992828369, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.019264448434114456, + "learning_rate": 0.0004528063409857554, + "loss": 3.6985023021698, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.01922295056283474, + "learning_rate": 0.0004525760908858209, + "loss": 3.651756763458252, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.01978331431746483, + "learning_rate": 0.00045234572027513174, + "loss": 3.679884910583496, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.01968599483370781, + "learning_rate": 0.00045211522933929317, + "loss": 3.6571812629699707, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.020741639658808708, + "learning_rate": 0.0004518846182640077, + "loss": 3.6505298614501953, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.01842017099261284, + "learning_rate": 0.0004516538872350744, + "loss": 3.6904940605163574, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.018004486337304115, + "learning_rate": 0.0004514230364383893, + "loss": 3.7282047271728516, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.018821008503437042, + "learning_rate": 0.0004511920660599444, + "loss": 3.668164014816284, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.018929854035377502, + "learning_rate": 0.0004509609762858287, + "loss": 3.6672980785369873, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.01979168877005577, + "learning_rate": 0.0004507297673022269, + "loss": 3.6684999465942383, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.01774384267628193, + "learning_rate": 0.0004504984392954199, + "loss": 3.6577749252319336, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01876000687479973, + "learning_rate": 0.00045026699245178444, + "loss": 3.6629252433776855, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.018197249621152878, + "learning_rate": 0.0004500354269577932, + "loss": 3.688983678817749, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.019496265798807144, + "learning_rate": 0.00044980374300001414, + "loss": 3.6602022647857666, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01825535111129284, + "learning_rate": 0.00044957194076511095, + "loss": 3.6889238357543945, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.017113137990236282, + "learning_rate": 0.0004493400204398426, + "loss": 3.684070587158203, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.017993126064538956, + "learning_rate": 0.00044910798221106283, + "loss": 3.675426721572876, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.018354708328843117, + "learning_rate": 0.00044887582626572105, + "loss": 3.675588607788086, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.016362832859158516, + "learning_rate": 0.0004486435527908606, + "loss": 3.6709303855895996, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01772182621061802, + "learning_rate": 0.0004484111619736205, + "loss": 3.6884102821350098, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.019271764904260635, + "learning_rate": 0.0004481786540012336, + "loss": 3.6468968391418457, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.019040528684854507, + "learning_rate": 0.00044794602906102747, + "loss": 3.690387010574341, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.019022610038518906, + "learning_rate": 0.00044771328734042367, + "loss": 3.689873218536377, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.018533391878008842, + "learning_rate": 0.00044748042902693815, + "loss": 3.6877284049987793, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.020825615152716637, + "learning_rate": 0.0004472474543081805, + "loss": 3.662257671356201, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.021880341693758965, + "learning_rate": 0.00044701436337185444, + "loss": 3.650700092315674, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.019974704831838608, + "learning_rate": 0.00044678115640575704, + "loss": 3.684177875518799, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.01903127320110798, + "learning_rate": 0.0004465478335977788, + "loss": 3.658148765563965, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.020999105647206306, + "learning_rate": 0.00044631439513590393, + "loss": 3.68747615814209, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.02138398587703705, + "learning_rate": 0.00044608084120820933, + "loss": 3.656355381011963, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.02216442860662937, + "learning_rate": 0.0004458471720028655, + "loss": 3.6860342025756836, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.019359121099114418, + "learning_rate": 0.0004456133877081352, + "loss": 3.654944658279419, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.020984770730137825, + "learning_rate": 0.0004453794885123744, + "loss": 3.6885194778442383, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.02423972077667713, + "learning_rate": 0.00044514547460403135, + "loss": 3.672712802886963, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.024544453248381615, + "learning_rate": 0.00044491134617164694, + "loss": 3.6484768390655518, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.060352325439453, + "eval_runtime": 57.9556, + "eval_samples_per_second": 42.136, + "eval_steps_per_second": 1.329, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.024072550237178802, + "learning_rate": 0.000444677103403854, + "loss": 3.5796468257904053, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.028701433911919594, + "learning_rate": 0.0004444427464893779, + "loss": 3.604318618774414, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.028119578957557678, + "learning_rate": 0.00044420827561703573, + "loss": 3.608759641647339, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.025335192680358887, + "learning_rate": 0.0004439736909757365, + "loss": 3.5760319232940674, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.0273929201066494, + "learning_rate": 0.0004437389927544805, + "loss": 3.559441089630127, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.026567714288830757, + "learning_rate": 0.0004435041811423603, + "loss": 3.626889228820801, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.027587885037064552, + "learning_rate": 0.0004432692563285591, + "loss": 3.6226272583007812, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.02935059368610382, + "learning_rate": 0.0004430342185023517, + "loss": 3.601921558380127, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.02982175536453724, + "learning_rate": 0.0004427990678531038, + "loss": 3.59737491607666, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.028570258989930153, + "learning_rate": 0.0004425638045702719, + "loss": 3.56337833404541, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.029890350997447968, + "learning_rate": 0.0004423284288434035, + "loss": 3.6294307708740234, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.03206814453005791, + "learning_rate": 0.00044209294086213657, + "loss": 3.590505599975586, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.03169345110654831, + "learning_rate": 0.00044185734081619957, + "loss": 3.581584930419922, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.03313758969306946, + "learning_rate": 0.000441621628895411, + "loss": 3.6088438034057617, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.03317603841423988, + "learning_rate": 0.00044138580528967986, + "loss": 3.6340456008911133, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.034738484770059586, + "learning_rate": 0.0004411498701890051, + "loss": 3.5814096927642822, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.036428093910217285, + "learning_rate": 0.0004409138237834751, + "loss": 3.5939526557922363, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.03293406963348389, + "learning_rate": 0.00044067766626326835, + "loss": 3.577897787094116, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.031030720099806786, + "learning_rate": 0.00044044139781865255, + "loss": 3.6077990531921387, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.03441236540675163, + "learning_rate": 0.00044020501863998516, + "loss": 3.591923713684082, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.03500965237617493, + "learning_rate": 0.00043996852891771236, + "loss": 3.6171927452087402, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.034443050622940063, + "learning_rate": 0.00043973192884236983, + "loss": 3.619724750518799, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.03435051813721657, + "learning_rate": 0.0004394952186045819, + "loss": 3.6443169116973877, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.03230702877044678, + "learning_rate": 0.00043925839839506167, + "loss": 3.5912671089172363, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.03232988342642784, + "learning_rate": 0.0004390214684046111, + "loss": 3.618985414505005, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.029298555105924606, + "learning_rate": 0.00043878442882412015, + "loss": 3.6166677474975586, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.026723450049757957, + "learning_rate": 0.00043854727984456746, + "loss": 3.61362886428833, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.028280874714255333, + "learning_rate": 0.0004383100216570197, + "loss": 3.6410951614379883, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.026700228452682495, + "learning_rate": 0.0004380726544526316, + "loss": 3.5804049968719482, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.023092713207006454, + "learning_rate": 0.00043783517842264534, + "loss": 3.595339298248291, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.024123897776007652, + "learning_rate": 0.0004375975937583914, + "loss": 3.647569417953491, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.02505572699010372, + "learning_rate": 0.0004373599006512871, + "loss": 3.6276652812957764, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.024523509666323662, + "learning_rate": 0.00043712209929283786, + "loss": 3.59525728225708, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.024351581931114197, + "learning_rate": 0.00043688418987463567, + "loss": 3.6081674098968506, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.02530614472925663, + "learning_rate": 0.00043664617258835995, + "loss": 3.6632542610168457, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.023438062518835068, + "learning_rate": 0.0004364080476257769, + "loss": 3.6345431804656982, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.021853657439351082, + "learning_rate": 0.0004361698151787396, + "loss": 3.598644256591797, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.022487422451376915, + "learning_rate": 0.0004359314754391874, + "loss": 3.5966577529907227, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.021353935822844505, + "learning_rate": 0.0004356930285991466, + "loss": 3.6328375339508057, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.021180452778935432, + "learning_rate": 0.00043545447485072926, + "loss": 3.6285812854766846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.020491832867264748, + "learning_rate": 0.000435215814386134, + "loss": 3.592944622039795, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.021603770554065704, + "learning_rate": 0.0004349770473976453, + "loss": 3.613507032394409, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.02146158367395401, + "learning_rate": 0.0004347381740776332, + "loss": 3.6409506797790527, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.02102433145046234, + "learning_rate": 0.0004344991946185538, + "loss": 3.6179330348968506, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.018684551119804382, + "learning_rate": 0.00043426010921294837, + "loss": 3.6385130882263184, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.018416227772831917, + "learning_rate": 0.0004340209180534438, + "loss": 3.642151355743408, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.01915144734084606, + "learning_rate": 0.000433781621332752, + "loss": 3.6089656352996826, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.018299665302038193, + "learning_rate": 0.0004335422192436703, + "loss": 3.6063547134399414, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.018962757661938667, + "learning_rate": 0.00043330271197908023, + "loss": 3.625917434692383, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.017437821254134178, + "learning_rate": 0.00043306309973194874, + "loss": 3.600015640258789, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.017913147807121277, + "learning_rate": 0.000432823382695327, + "loss": 3.6428744792938232, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.018343618139624596, + "learning_rate": 0.0004325835610623508, + "loss": 3.635901927947998, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.020372850820422173, + "learning_rate": 0.00043234363502623994, + "loss": 3.620181083679199, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.02037600241601467, + "learning_rate": 0.0004321036047802985, + "loss": 3.61129093170166, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.021827151998877525, + "learning_rate": 0.00043186347051791466, + "loss": 3.5999555587768555, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.021806657314300537, + "learning_rate": 0.0004316232324325602, + "loss": 3.642509937286377, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02050699107348919, + "learning_rate": 0.0004313828907177906, + "loss": 3.597583293914795, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.020436758175492287, + "learning_rate": 0.0004311424455672448, + "loss": 3.6079912185668945, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.019341666251420975, + "learning_rate": 0.0004309018971746451, + "loss": 3.6196017265319824, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.022348817437887192, + "learning_rate": 0.000430661245733797, + "loss": 3.619025468826294, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.019361121580004692, + "learning_rate": 0.00043042049143858914, + "loss": 3.657182455062866, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.019258935004472733, + "learning_rate": 0.0004301796344829928, + "loss": 3.6408963203430176, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.020622272044420242, + "learning_rate": 0.00042993867506106205, + "loss": 3.6247220039367676, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.019510755315423012, + "learning_rate": 0.0004296976133669336, + "loss": 3.6264584064483643, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.02078665979206562, + "learning_rate": 0.0004294564495948266, + "loss": 3.6364388465881348, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01791800744831562, + "learning_rate": 0.0004292151839390421, + "loss": 3.6029374599456787, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.01922667771577835, + "learning_rate": 0.0004289738165939636, + "loss": 3.582937717437744, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.018473893404006958, + "learning_rate": 0.00042873234775405654, + "loss": 3.6141819953918457, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01846298575401306, + "learning_rate": 0.00042849077761386784, + "loss": 3.638786554336548, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.01876973733305931, + "learning_rate": 0.0004282491063680263, + "loss": 3.617130756378174, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.017995499074459076, + "learning_rate": 0.00042800733421124207, + "loss": 3.645305633544922, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.02105429396033287, + "learning_rate": 0.00042776546133830646, + "loss": 3.6444876194000244, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.049062252044678, + "eval_runtime": 58.0113, + "eval_samples_per_second": 42.095, + "eval_steps_per_second": 1.327, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.02128119394183159, + "learning_rate": 0.00042752348794409227, + "loss": 3.524954080581665, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.028249500319361687, + "learning_rate": 0.00042728141422355287, + "loss": 3.5311479568481445, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.03145391866564751, + "learning_rate": 0.0004270392403717229, + "loss": 3.559330940246582, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.02994118444621563, + "learning_rate": 0.00042679696658371743, + "loss": 3.5257620811462402, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.027448246255517006, + "learning_rate": 0.00042655459305473195, + "loss": 3.5242764949798584, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.0296134352684021, + "learning_rate": 0.0004263121199800425, + "loss": 3.5317091941833496, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.029576804488897324, + "learning_rate": 0.0004260695475550054, + "loss": 3.5340428352355957, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.02941269613802433, + "learning_rate": 0.0004258268759750566, + "loss": 3.5372023582458496, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.03126693144440651, + "learning_rate": 0.0004255841054357124, + "loss": 3.5437135696411133, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.03335367515683174, + "learning_rate": 0.0004253412361325686, + "loss": 3.5598273277282715, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.03028946928679943, + "learning_rate": 0.0004250982682613006, + "loss": 3.5190839767456055, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.028629133477807045, + "learning_rate": 0.00042485520201766315, + "loss": 3.5138821601867676, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.027070026844739914, + "learning_rate": 0.0004246120375974905, + "loss": 3.530958414077759, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.028837760910391808, + "learning_rate": 0.0004243687751966957, + "loss": 3.5233962535858154, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.029884297400712967, + "learning_rate": 0.0004241254150112708, + "loss": 3.5639758110046387, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.024223247542977333, + "learning_rate": 0.000423881957237287, + "loss": 3.5555152893066406, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.026992836967110634, + "learning_rate": 0.0004236384020708935, + "loss": 3.540196418762207, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.02818330004811287, + "learning_rate": 0.0004233947497083185, + "loss": 3.5546071529388428, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.026401042938232422, + "learning_rate": 0.0004231510003458682, + "loss": 3.5556247234344482, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.026020489633083344, + "learning_rate": 0.0004229071541799272, + "loss": 3.568004608154297, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.02803782932460308, + "learning_rate": 0.0004226632114069578, + "loss": 3.5828235149383545, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.02737978659570217, + "learning_rate": 0.0004224191722235004, + "loss": 3.5642499923706055, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.02406371757388115, + "learning_rate": 0.00042217503682617286, + "loss": 3.5530993938446045, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.02230692468583584, + "learning_rate": 0.0004219308054116706, + "loss": 3.5573501586914062, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.021352406591176987, + "learning_rate": 0.0004216864781767667, + "loss": 3.5526974201202393, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.02227999083697796, + "learning_rate": 0.00042144205531831103, + "loss": 3.5866928100585938, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.024190247058868408, + "learning_rate": 0.0004211975370332308, + "loss": 3.517117977142334, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.023196619004011154, + "learning_rate": 0.00042095292351852976, + "loss": 3.4826502799987793, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.02388714998960495, + "learning_rate": 0.0004207082149712888, + "loss": 3.5588626861572266, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.024692339822649956, + "learning_rate": 0.0004204634115886651, + "loss": 3.54828143119812, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.023078206926584244, + "learning_rate": 0.0004202185135678924, + "loss": 3.5679030418395996, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.020665735006332397, + "learning_rate": 0.00041997352110628033, + "loss": 3.5500283241271973, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.02022322081029415, + "learning_rate": 0.00041972843440121527, + "loss": 3.592104911804199, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.021137656643986702, + "learning_rate": 0.00041948325365015894, + "loss": 3.583833694458008, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.021026931703090668, + "learning_rate": 0.0004192379790506491, + "loss": 3.556403875350952, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.020423648878932, + "learning_rate": 0.00041899261080029904, + "loss": 3.5839173793792725, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.020935002714395523, + "learning_rate": 0.00041874714909679754, + "loss": 3.5794217586517334, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.020799759775400162, + "learning_rate": 0.00041850159413790863, + "loss": 3.6033859252929688, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.019480884075164795, + "learning_rate": 0.0004182559461214715, + "loss": 3.580097198486328, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.019300777465105057, + "learning_rate": 0.0004180102052454004, + "loss": 3.550474166870117, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.020265772938728333, + "learning_rate": 0.00041776437170768427, + "loss": 3.5711655616760254, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.019972780719399452, + "learning_rate": 0.0004175184457063868, + "loss": 3.5636885166168213, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.019635623320937157, + "learning_rate": 0.000417272427439646, + "loss": 3.5851311683654785, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.021451901644468307, + "learning_rate": 0.00041702631710567443, + "loss": 3.5956759452819824, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.019456414505839348, + "learning_rate": 0.00041678011490275875, + "loss": 3.595731258392334, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.019757689908146858, + "learning_rate": 0.00041653382102925957, + "loss": 3.569652557373047, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.020571604371070862, + "learning_rate": 0.00041628743568361147, + "loss": 3.6007747650146484, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.02153848111629486, + "learning_rate": 0.00041604095906432265, + "loss": 3.5915260314941406, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.020875850692391396, + "learning_rate": 0.00041579439136997493, + "loss": 3.5582690238952637, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.02089419960975647, + "learning_rate": 0.00041554773279922337, + "loss": 3.5756618976593018, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01961098238825798, + "learning_rate": 0.0004153009835507963, + "loss": 3.5918776988983154, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.021075716242194176, + "learning_rate": 0.0004150541438234952, + "loss": 3.5342202186584473, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.020745476707816124, + "learning_rate": 0.00041480721381619435, + "loss": 3.5976028442382812, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.02001352794468403, + "learning_rate": 0.00041456019372784086, + "loss": 3.5686182975769043, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.021684160456061363, + "learning_rate": 0.0004143130837574543, + "loss": 3.553445339202881, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.022551920264959335, + "learning_rate": 0.0004140658841041267, + "loss": 3.5712709426879883, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.022758224979043007, + "learning_rate": 0.0004138185949670224, + "loss": 3.6101667881011963, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.020714636892080307, + "learning_rate": 0.00041357121654537766, + "loss": 3.5740506649017334, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.022694187238812447, + "learning_rate": 0.00041332374903850104, + "loss": 3.5663979053497314, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.021664340049028397, + "learning_rate": 0.00041307619264577234, + "loss": 3.5830888748168945, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.019072270020842552, + "learning_rate": 0.0004128285475666436, + "loss": 3.5861973762512207, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.020153764635324478, + "learning_rate": 0.0004125808140006378, + "loss": 3.566444158554077, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019666558131575584, + "learning_rate": 0.0004123329921473494, + "loss": 3.5789592266082764, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.02101624570786953, + "learning_rate": 0.0004120850822064439, + "loss": 3.551722526550293, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.01951020397245884, + "learning_rate": 0.00041183708437765815, + "loss": 3.6069014072418213, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.01985476352274418, + "learning_rate": 0.00041158899886079925, + "loss": 3.5944924354553223, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02195335365831852, + "learning_rate": 0.00041134082585574534, + "loss": 3.5828046798706055, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02105172537267208, + "learning_rate": 0.00041109256556244497, + "loss": 3.5830953121185303, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.01888827420771122, + "learning_rate": 0.00041084421818091685, + "loss": 3.5656847953796387, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.017867228016257286, + "learning_rate": 0.0004105957839112501, + "loss": 3.5792243480682373, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.018577905371785164, + "learning_rate": 0.0004103472629536036, + "loss": 3.56992506980896, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.020045632496476173, + "learning_rate": 0.0004100986555082063, + "loss": 3.590876579284668, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.044339656829834, + "eval_runtime": 57.1435, + "eval_samples_per_second": 42.735, + "eval_steps_per_second": 1.347, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.024135645478963852, + "learning_rate": 0.0004098499617753567, + "loss": 3.4856467247009277, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.031053155660629272, + "learning_rate": 0.00040960118195542273, + "loss": 3.455773115158081, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.035190410912036896, + "learning_rate": 0.00040935231624884204, + "loss": 3.506619930267334, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.03611094877123833, + "learning_rate": 0.0004091033648561211, + "loss": 3.5227696895599365, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.034121073782444, + "learning_rate": 0.00040885432797783543, + "loss": 3.4970145225524902, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.03629382699728012, + "learning_rate": 0.0004086052058146297, + "loss": 3.506425619125366, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.03709045797586441, + "learning_rate": 0.00040835599856721725, + "loss": 3.499274969100952, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.034826792776584625, + "learning_rate": 0.00040810670643637954, + "loss": 3.5029163360595703, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.03344690054655075, + "learning_rate": 0.000407857329622967, + "loss": 3.4816911220550537, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.03136949986219406, + "learning_rate": 0.00040760786832789786, + "loss": 3.5154659748077393, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.030533097684383392, + "learning_rate": 0.0004073583227521588, + "loss": 3.509666919708252, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.03298318013548851, + "learning_rate": 0.0004071086930968039, + "loss": 3.5096616744995117, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.031186232343316078, + "learning_rate": 0.00040685897956295545, + "loss": 3.4752590656280518, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.028008371591567993, + "learning_rate": 0.0004066091823518031, + "loss": 3.498826503753662, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.028206519782543182, + "learning_rate": 0.00040635930166460385, + "loss": 3.5281882286071777, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.024417119100689888, + "learning_rate": 0.00040610933770268226, + "loss": 3.506791830062866, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.025321070104837418, + "learning_rate": 0.00040585929066742964, + "loss": 3.494256019592285, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.023430505767464638, + "learning_rate": 0.00040560916076030435, + "loss": 3.533010482788086, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.023371374234557152, + "learning_rate": 0.00040535894818283154, + "loss": 3.5197086334228516, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.02414027415215969, + "learning_rate": 0.0004051086531366031, + "loss": 3.4707889556884766, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.026652060449123383, + "learning_rate": 0.000404858275823277, + "loss": 3.498159885406494, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.024486590176820755, + "learning_rate": 0.000404607816444578, + "loss": 3.50034236907959, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.023153992369771004, + "learning_rate": 0.00040435727520229644, + "loss": 3.499504804611206, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.02222427912056446, + "learning_rate": 0.0004041066522982891, + "loss": 3.4953579902648926, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.023033641278743744, + "learning_rate": 0.000403855947934478, + "loss": 3.5181496143341064, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.02141181193292141, + "learning_rate": 0.00040360516231285155, + "loss": 3.552860736846924, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.023448696359992027, + "learning_rate": 0.00040335429563546286, + "loss": 3.5153324604034424, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.022410666570067406, + "learning_rate": 0.0004031033481044308, + "loss": 3.4963512420654297, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.020326802507042885, + "learning_rate": 0.00040285231992193924, + "loss": 3.528319835662842, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021117394790053368, + "learning_rate": 0.0004026012112902372, + "loss": 3.5192837715148926, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.021317072212696075, + "learning_rate": 0.0004023500224116381, + "loss": 3.5081021785736084, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.021116144955158234, + "learning_rate": 0.00040209875348852037, + "loss": 3.5167174339294434, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019834019243717194, + "learning_rate": 0.00040184740472332705, + "loss": 3.511960983276367, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.02089378982782364, + "learning_rate": 0.000401595976318565, + "loss": 3.493330955505371, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.021129710599780083, + "learning_rate": 0.00040134446847680584, + "loss": 3.544156551361084, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.021163299679756165, + "learning_rate": 0.0004010928814006846, + "loss": 3.5261003971099854, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.02105306088924408, + "learning_rate": 0.0004008412152929007, + "loss": 3.543973445892334, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.02150859497487545, + "learning_rate": 0.0004005894703562168, + "loss": 3.487217426300049, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.021666452288627625, + "learning_rate": 0.0004003376467934593, + "loss": 3.4933671951293945, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.02180723287165165, + "learning_rate": 0.0004000857448075178, + "loss": 3.55198073387146, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.02041625790297985, + "learning_rate": 0.0003998337646013455, + "loss": 3.529571533203125, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.022977063432335854, + "learning_rate": 0.00039958170637795806, + "loss": 3.526432752609253, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.022725161164999008, + "learning_rate": 0.0003993295703404341, + "loss": 3.5258829593658447, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.021337240934371948, + "learning_rate": 0.00039907735669191544, + "loss": 3.4897351264953613, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.020783131942152977, + "learning_rate": 0.00039882506563560573, + "loss": 3.5497589111328125, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.022259317338466644, + "learning_rate": 0.0003985726973747715, + "loss": 3.5225064754486084, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.020283272489905357, + "learning_rate": 0.00039832025211274113, + "loss": 3.5453078746795654, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.020656265318393707, + "learning_rate": 0.00039806773005290544, + "loss": 3.533703327178955, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.02104976586997509, + "learning_rate": 0.0003978151313987168, + "loss": 3.4885146617889404, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.021411150693893433, + "learning_rate": 0.00039756245635368933, + "loss": 3.533268690109253, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.021906673908233643, + "learning_rate": 0.00039730970512139876, + "loss": 3.5483224391937256, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.02415728010237217, + "learning_rate": 0.0003970568779054822, + "loss": 3.5084290504455566, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.023435061797499657, + "learning_rate": 0.000396803974909638, + "loss": 3.5498428344726562, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.02183385379612446, + "learning_rate": 0.0003965509963376255, + "loss": 3.484086751937866, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.022856134921312332, + "learning_rate": 0.000396297942393265, + "loss": 3.5339231491088867, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.021679896861314774, + "learning_rate": 0.0003960448132804375, + "loss": 3.547191619873047, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.021262124180793762, + "learning_rate": 0.0003957916092030845, + "loss": 3.5263822078704834, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.02265402302145958, + "learning_rate": 0.00039553833036520803, + "loss": 3.5361135005950928, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.020080721005797386, + "learning_rate": 0.0003952849769708702, + "loss": 3.5247459411621094, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.02187187783420086, + "learning_rate": 0.0003950315492241932, + "loss": 3.54129695892334, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.02417500875890255, + "learning_rate": 0.0003947780473293593, + "loss": 3.598796844482422, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.02416270412504673, + "learning_rate": 0.00039452447149061054, + "loss": 3.5207936763763428, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.021958643570542336, + "learning_rate": 0.0003942708219122481, + "loss": 3.5395452976226807, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.021813420578837395, + "learning_rate": 0.000394017098798633, + "loss": 3.5469255447387695, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020607495680451393, + "learning_rate": 0.0003937633023541854, + "loss": 3.554851531982422, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.02091527171432972, + "learning_rate": 0.0003935094327833845, + "loss": 3.492576837539673, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.021089257672429085, + "learning_rate": 0.0003932554902907683, + "loss": 3.5154709815979004, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.019659532234072685, + "learning_rate": 0.0003930014750809338, + "loss": 3.5444283485412598, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01896163448691368, + "learning_rate": 0.0003927473873585365, + "loss": 3.589625358581543, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020288823172450066, + "learning_rate": 0.0003924932273282903, + "loss": 3.522411823272705, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.020222362130880356, + "learning_rate": 0.00039223899519496723, + "loss": 3.5244359970092773, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.022349944338202477, + "learning_rate": 0.00039198469116339757, + "loss": 3.5362396240234375, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.047028064727783, + "eval_runtime": 62.8948, + "eval_samples_per_second": 38.827, + "eval_steps_per_second": 1.224, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.02202913723886013, + "learning_rate": 0.0003917303154384694, + "loss": 3.4249138832092285, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.03153969347476959, + "learning_rate": 0.00039147586822512885, + "loss": 3.4436073303222656, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.03747870773077011, + "learning_rate": 0.0003912213497283793, + "loss": 3.4879579544067383, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.036144088953733444, + "learning_rate": 0.0003909667601532819, + "loss": 3.419151782989502, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.03983336314558983, + "learning_rate": 0.00039071209970495445, + "loss": 3.4331436157226562, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.045912645757198334, + "learning_rate": 0.00039045736858857273, + "loss": 3.4304769039154053, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.04005753993988037, + "learning_rate": 0.0003902025670093687, + "loss": 3.478600025177002, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.03488151356577873, + "learning_rate": 0.00038994769517263156, + "loss": 3.4650232791900635, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.03603382781147957, + "learning_rate": 0.0003896927532837069, + "loss": 3.449807643890381, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.03817906603217125, + "learning_rate": 0.0003894377415479967, + "loss": 3.4556589126586914, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.03523671627044678, + "learning_rate": 0.00038918266017095936, + "loss": 3.477426767349243, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.03284068778157234, + "learning_rate": 0.0003889275093581096, + "loss": 3.478672981262207, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.02992352657020092, + "learning_rate": 0.00038867228931501737, + "loss": 3.4810047149658203, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.02858242578804493, + "learning_rate": 0.0003884170002473091, + "loss": 3.467811107635498, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.027912171557545662, + "learning_rate": 0.0003881616423606666, + "loss": 3.4326171875, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.028351375833153725, + "learning_rate": 0.00038790621586082697, + "loss": 3.4533729553222656, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.027947887778282166, + "learning_rate": 0.0003876507209535829, + "loss": 3.420295476913452, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.028510035946965218, + "learning_rate": 0.00038739515784478186, + "loss": 3.438023328781128, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.027707934379577637, + "learning_rate": 0.00038713952674032654, + "loss": 3.4844181537628174, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.0262764822691679, + "learning_rate": 0.0003868838278461743, + "loss": 3.4363749027252197, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.02713681012392044, + "learning_rate": 0.0003866280613683371, + "loss": 3.4501900672912598, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.0299485195428133, + "learning_rate": 0.00038637222751288126, + "loss": 3.461942672729492, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.027047986164689064, + "learning_rate": 0.0003861163264859277, + "loss": 3.454293727874756, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.02436736971139908, + "learning_rate": 0.0003858603584936511, + "loss": 3.4663405418395996, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.02601189725100994, + "learning_rate": 0.0003856043237422803, + "loss": 3.502030849456787, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.02463468722999096, + "learning_rate": 0.0003853482224380978, + "loss": 3.462491035461426, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.021793190389871597, + "learning_rate": 0.00038509205478744, + "loss": 3.4476377964019775, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.023809107020497322, + "learning_rate": 0.0003848358209966963, + "loss": 3.4701178073883057, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.02557182125747204, + "learning_rate": 0.0003845795212723099, + "loss": 3.4932539463043213, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.023424552753567696, + "learning_rate": 0.00038432315582077664, + "loss": 3.4779105186462402, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.02195148915052414, + "learning_rate": 0.0003840667248486456, + "loss": 3.470506191253662, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.022897250950336456, + "learning_rate": 0.0003838102285625186, + "loss": 3.48982572555542, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.023033805191516876, + "learning_rate": 0.00038355366716905006, + "loss": 3.4845101833343506, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.022774485871195793, + "learning_rate": 0.0003832970408749467, + "loss": 3.5060324668884277, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.02193611115217209, + "learning_rate": 0.00038304034988696816, + "loss": 3.511518955230713, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.02130570448935032, + "learning_rate": 0.00038278359441192516, + "loss": 3.468502998352051, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.020402755588293076, + "learning_rate": 0.00038252677465668136, + "loss": 3.4786908626556396, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.022664198651909828, + "learning_rate": 0.00038226989082815156, + "loss": 3.479928970336914, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.02454056777060032, + "learning_rate": 0.00038201294313330245, + "loss": 3.479947566986084, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.022096732631325722, + "learning_rate": 0.00038175593177915227, + "loss": 3.49478816986084, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.021499406546354294, + "learning_rate": 0.0003814988569727704, + "loss": 3.494933605194092, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.022325703874230385, + "learning_rate": 0.0003812417189212773, + "loss": 3.5041463375091553, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.020674455910921097, + "learning_rate": 0.00038098451783184447, + "loss": 3.472404718399048, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.02131054364144802, + "learning_rate": 0.00038072725391169416, + "loss": 3.497795343399048, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.02126218006014824, + "learning_rate": 0.0003804699273680994, + "loss": 3.483281373977661, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.0217830128967762, + "learning_rate": 0.0003802125384083834, + "loss": 3.4704298973083496, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.02156917378306389, + "learning_rate": 0.0003799550872399197, + "loss": 3.4994399547576904, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.02047055773437023, + "learning_rate": 0.0003796975740701323, + "loss": 3.4316084384918213, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.019766805693507195, + "learning_rate": 0.0003794399991064948, + "loss": 3.452223777770996, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.020028885453939438, + "learning_rate": 0.00037918236255653074, + "loss": 3.485278844833374, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.020312055945396423, + "learning_rate": 0.00037892466462781306, + "loss": 3.4789156913757324, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.020725280046463013, + "learning_rate": 0.00037866690552796446, + "loss": 3.495528221130371, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.02088077738881111, + "learning_rate": 0.00037840908546465696, + "loss": 3.4972569942474365, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.021010037511587143, + "learning_rate": 0.0003781512046456111, + "loss": 3.473562240600586, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.019641095772385597, + "learning_rate": 0.0003778932632785972, + "loss": 3.4782915115356445, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.02063329517841339, + "learning_rate": 0.00037763526157143376, + "loss": 3.4681782722473145, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.0213928259909153, + "learning_rate": 0.00037737719973198825, + "loss": 3.472263813018799, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.02135022170841694, + "learning_rate": 0.0003771190779681762, + "loss": 3.502300262451172, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.02053583413362503, + "learning_rate": 0.0003768608964879619, + "loss": 3.489243745803833, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.021575238555669785, + "learning_rate": 0.00037660265549935724, + "loss": 3.491074562072754, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.022335771471261978, + "learning_rate": 0.00037634435521042257, + "loss": 3.5298843383789062, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.021198464557528496, + "learning_rate": 0.0003760859958292656, + "loss": 3.5156898498535156, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.020986037328839302, + "learning_rate": 0.000375827577564042, + "loss": 3.4574475288391113, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.022639811038970947, + "learning_rate": 0.0003755691006229545, + "loss": 3.4952611923217773, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.019711066037416458, + "learning_rate": 0.0003753105652142534, + "loss": 3.531296730041504, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.02058061584830284, + "learning_rate": 0.000375051971546236, + "loss": 3.5340657234191895, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.019849255681037903, + "learning_rate": 0.0003747933198272465, + "loss": 3.5028295516967773, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01954430714249611, + "learning_rate": 0.00037453461026567604, + "loss": 3.51088809967041, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.019234873354434967, + "learning_rate": 0.00037427584306996196, + "loss": 3.472560167312622, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.021085912361741066, + "learning_rate": 0.0003740170184485888, + "loss": 3.468876838684082, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.020300406962633133, + "learning_rate": 0.0003737581366100864, + "loss": 3.4903886318206787, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02380351349711418, + "learning_rate": 0.0003734991977630315, + "loss": 3.5069704055786133, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.048111915588379, + "eval_runtime": 58.312, + "eval_samples_per_second": 41.878, + "eval_steps_per_second": 1.32, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02379327267408371, + "learning_rate": 0.0003732402021160463, + "loss": 3.396376132965088, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.03204222396016121, + "learning_rate": 0.00037298114987779885, + "loss": 3.4102582931518555, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.034993767738342285, + "learning_rate": 0.000372722041257003, + "loss": 3.385697364807129, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.03387003391981125, + "learning_rate": 0.00037246287646241783, + "loss": 3.4120078086853027, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.03282478079199791, + "learning_rate": 0.0003722036557028478, + "loss": 3.404122829437256, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.03053346462547779, + "learning_rate": 0.0003719443791871422, + "loss": 3.368051767349243, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.029722390696406364, + "learning_rate": 0.0003716850471241958, + "loss": 3.4066247940063477, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.0287226140499115, + "learning_rate": 0.0003714256597229474, + "loss": 3.4212646484375, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.028156550601124763, + "learning_rate": 0.0003711662171923809, + "loss": 3.4192593097686768, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.03021939843893051, + "learning_rate": 0.0003709067197415244, + "loss": 3.3973188400268555, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.03314025700092316, + "learning_rate": 0.00037064716757945036, + "loss": 3.422978401184082, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.03156686946749687, + "learning_rate": 0.0003703875609152753, + "loss": 3.4052343368530273, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.03357478231191635, + "learning_rate": 0.0003701278999581595, + "loss": 3.383730411529541, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.03261081501841545, + "learning_rate": 0.0003698681849173073, + "loss": 3.421250820159912, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.03252755105495453, + "learning_rate": 0.0003696084160019662, + "loss": 3.39884352684021, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.03177274018526077, + "learning_rate": 0.0003693485934214274, + "loss": 3.4175422191619873, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.028845487162470818, + "learning_rate": 0.0003690887173850253, + "loss": 3.3908143043518066, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.026045162230730057, + "learning_rate": 0.0003688287881021374, + "loss": 3.4417710304260254, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.027937524020671844, + "learning_rate": 0.00036856880578218376, + "loss": 3.410548210144043, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.025780271738767624, + "learning_rate": 0.00036830877063462786, + "loss": 3.4069018363952637, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.026093080639839172, + "learning_rate": 0.0003680486828689749, + "loss": 3.4199600219726562, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.025765180587768555, + "learning_rate": 0.00036778854269477315, + "loss": 3.442903995513916, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.022773101925849915, + "learning_rate": 0.0003675283503216127, + "loss": 3.394232749938965, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.025135623291134834, + "learning_rate": 0.000367268105959126, + "loss": 3.4466171264648438, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.02491890639066696, + "learning_rate": 0.00036700780981698705, + "loss": 3.418689727783203, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.02521841786801815, + "learning_rate": 0.0003667474621049119, + "loss": 3.4213061332702637, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.025186199694871902, + "learning_rate": 0.00036648706303265795, + "loss": 3.437631130218506, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.022097431123256683, + "learning_rate": 0.000366226612810024, + "loss": 3.4014902114868164, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.024009618908166885, + "learning_rate": 0.00036596611164685025, + "loss": 3.4292187690734863, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.022758696228265762, + "learning_rate": 0.0003657055597530175, + "loss": 3.4105381965637207, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.02447321265935898, + "learning_rate": 0.00036544495733844797, + "loss": 3.4196958541870117, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.025843804702162743, + "learning_rate": 0.000365184304613104, + "loss": 3.4343783855438232, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.022433951497077942, + "learning_rate": 0.0003649236017869892, + "loss": 3.440580368041992, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.02250702679157257, + "learning_rate": 0.00036466284907014685, + "loss": 3.4255261421203613, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.022953296080231667, + "learning_rate": 0.0003644020466726608, + "loss": 3.442258358001709, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.021977661177515984, + "learning_rate": 0.00036414119480465473, + "loss": 3.4058687686920166, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.021751178428530693, + "learning_rate": 0.0003638802936762925, + "loss": 3.4211649894714355, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.022764205932617188, + "learning_rate": 0.0003636193434977771, + "loss": 3.424694061279297, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.024827787652611732, + "learning_rate": 0.0003633583444793516, + "loss": 3.458076238632202, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.021888602524995804, + "learning_rate": 0.00036309729683129814, + "loss": 3.4509153366088867, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.021961702033877373, + "learning_rate": 0.00036283620076393814, + "loss": 3.4147253036499023, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.02235778421163559, + "learning_rate": 0.000362575056487632, + "loss": 3.4392640590667725, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.021714260801672935, + "learning_rate": 0.00036231386421277877, + "loss": 3.444026470184326, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.020996905863285065, + "learning_rate": 0.0003620526241498167, + "loss": 3.4168434143066406, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.020528919994831085, + "learning_rate": 0.0003617913365092218, + "loss": 3.4831130504608154, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.021056141704320908, + "learning_rate": 0.0003615300015015091, + "loss": 3.430905342102051, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.021655110642313957, + "learning_rate": 0.0003612686193372312, + "loss": 3.4588241577148438, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.02344612590968609, + "learning_rate": 0.00036100719022697924, + "loss": 3.4202680587768555, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.02264723740518093, + "learning_rate": 0.0003607457143813819, + "loss": 3.4669015407562256, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.02377854660153389, + "learning_rate": 0.00036048419201110553, + "loss": 3.5059216022491455, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.02299923449754715, + "learning_rate": 0.00036022262332685383, + "loss": 3.4796524047851562, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.021110547706484795, + "learning_rate": 0.0003599610085393681, + "loss": 3.4342408180236816, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.022060122340917587, + "learning_rate": 0.0003596993478594267, + "loss": 3.426923990249634, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.021077830344438553, + "learning_rate": 0.0003594376414978447, + "loss": 3.4636270999908447, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.02209082990884781, + "learning_rate": 0.0003591758896654745, + "loss": 3.441979169845581, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.022408071905374527, + "learning_rate": 0.0003589140925732045, + "loss": 3.490835666656494, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.022462666034698486, + "learning_rate": 0.0003586522504319602, + "loss": 3.452770233154297, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.021295800805091858, + "learning_rate": 0.0003583903634527029, + "loss": 3.4377408027648926, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.020457757636904716, + "learning_rate": 0.0003581284318464302, + "loss": 3.4191250801086426, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.021163633093237877, + "learning_rate": 0.00035786645582417564, + "loss": 3.4797091484069824, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.021495461463928223, + "learning_rate": 0.00035760443559700863, + "loss": 3.4763145446777344, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.021050071343779564, + "learning_rate": 0.00035734237137603417, + "loss": 3.4687209129333496, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.02150929905474186, + "learning_rate": 0.0003570802633723927, + "loss": 3.4819254875183105, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.0212967898696661, + "learning_rate": 0.0003568181117972597, + "loss": 3.490407705307007, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.02123691327869892, + "learning_rate": 0.00035655591686184626, + "loss": 3.4418535232543945, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.02193271741271019, + "learning_rate": 0.0003562936787773979, + "loss": 3.446892738342285, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.02065582573413849, + "learning_rate": 0.000356031397755195, + "loss": 3.4308691024780273, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.019487040117383003, + "learning_rate": 0.00035576907400655303, + "loss": 3.487239122390747, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.021190891042351723, + "learning_rate": 0.00035550670774282106, + "loss": 3.4635558128356934, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.0209872554987669, + "learning_rate": 0.0003552442991753831, + "loss": 3.443793773651123, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.02035645768046379, + "learning_rate": 0.00035498184851565697, + "loss": 3.473198652267456, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.024002201855182648, + "learning_rate": 0.0003547193559750944, + "loss": 3.4244751930236816, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.057085990905762, + "eval_runtime": 62.4735, + "eval_samples_per_second": 39.089, + "eval_steps_per_second": 1.233, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.02450242079794407, + "learning_rate": 0.00035445682176518087, + "loss": 3.361237049102783, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.030813416466116905, + "learning_rate": 0.00035419424609743546, + "loss": 3.365708351135254, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.03279578685760498, + "learning_rate": 0.00035393162918341065, + "loss": 3.325385570526123, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.03293970227241516, + "learning_rate": 0.00035366897123469213, + "loss": 3.356811285018921, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.033190157264471054, + "learning_rate": 0.0003534062724628986, + "loss": 3.406217098236084, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.033370476216077805, + "learning_rate": 0.00035314353307968195, + "loss": 3.3579213619232178, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.03351330757141113, + "learning_rate": 0.00035288075329672634, + "loss": 3.3755106925964355, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.034596871584653854, + "learning_rate": 0.0003526179333257487, + "loss": 3.3576536178588867, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.03177076578140259, + "learning_rate": 0.0003523550733784984, + "loss": 3.3123257160186768, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.0289607010781765, + "learning_rate": 0.00035209217366675706, + "loss": 3.370959997177124, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.03265267238020897, + "learning_rate": 0.000351829234402338, + "loss": 3.332608938217163, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.03397805988788605, + "learning_rate": 0.0003515662557970867, + "loss": 3.364213228225708, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.03100395016372204, + "learning_rate": 0.0003513032380628804, + "loss": 3.388005256652832, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.03224753588438034, + "learning_rate": 0.0003510401814116277, + "loss": 3.3267595767974854, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.03034576214849949, + "learning_rate": 0.0003507770860552684, + "loss": 3.3954129219055176, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.026316817849874496, + "learning_rate": 0.00035051395220577397, + "loss": 3.3543832302093506, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.027714375406503677, + "learning_rate": 0.0003502507800751464, + "loss": 3.4190099239349365, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.028731998056173325, + "learning_rate": 0.0003499875698754187, + "loss": 3.342031955718994, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02845810353755951, + "learning_rate": 0.00034972432181865467, + "loss": 3.356130599975586, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.02842753194272518, + "learning_rate": 0.00034946103611694854, + "loss": 3.3709869384765625, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.02942183054983616, + "learning_rate": 0.00034919771298242477, + "loss": 3.373588800430298, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.02958574891090393, + "learning_rate": 0.00034893435262723814, + "loss": 3.377514600753784, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.029524607583880424, + "learning_rate": 0.00034867095526357325, + "loss": 3.398663282394409, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.028092598542571068, + "learning_rate": 0.0003484075211036446, + "loss": 3.3742518424987793, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.026652829721570015, + "learning_rate": 0.0003481440503596964, + "loss": 3.387667179107666, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.024675756692886353, + "learning_rate": 0.0003478805432440021, + "loss": 3.3737070560455322, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.02491692826151848, + "learning_rate": 0.0003476169999688648, + "loss": 3.359217643737793, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.024601681157946587, + "learning_rate": 0.00034735342074661654, + "loss": 3.3902688026428223, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.02313913404941559, + "learning_rate": 0.0003470898057896183, + "loss": 3.3700339794158936, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.02418898604810238, + "learning_rate": 0.0003468261553102599, + "loss": 3.4265692234039307, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.023382030427455902, + "learning_rate": 0.00034656246952095984, + "loss": 3.378389835357666, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.02340470254421234, + "learning_rate": 0.0003462987486341648, + "loss": 3.392937421798706, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.022166339680552483, + "learning_rate": 0.00034603499286235006, + "loss": 3.39414119720459, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.022175129503011703, + "learning_rate": 0.0003457712024180188, + "loss": 3.407447099685669, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.02368207648396492, + "learning_rate": 0.0003455073775137025, + "loss": 3.415076971054077, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.02263982594013214, + "learning_rate": 0.0003452435183619598, + "loss": 3.383986473083496, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.02147712931036949, + "learning_rate": 0.0003449796251753773, + "loss": 3.403963565826416, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.022987166419625282, + "learning_rate": 0.00034471569816656915, + "loss": 3.3883113861083984, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.02240179106593132, + "learning_rate": 0.00034445173754817646, + "loss": 3.4349911212921143, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.02206592820584774, + "learning_rate": 0.00034418774353286747, + "loss": 3.426473617553711, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.022650036960840225, + "learning_rate": 0.0003439237163333375, + "loss": 3.4251773357391357, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.023399339988827705, + "learning_rate": 0.0003436596561623084, + "loss": 3.371610641479492, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.02192695252597332, + "learning_rate": 0.0003433955632325288, + "loss": 3.387253761291504, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.02241935208439827, + "learning_rate": 0.00034313143775677353, + "loss": 3.434523105621338, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.023202672600746155, + "learning_rate": 0.00034286727994784367, + "loss": 3.369771957397461, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.021333014592528343, + "learning_rate": 0.0003426030900185665, + "loss": 3.439004421234131, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.023988714441657066, + "learning_rate": 0.0003423388681817949, + "loss": 3.396838665008545, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.023885777220129967, + "learning_rate": 0.00034207461465040793, + "loss": 3.4114322662353516, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.02263713628053665, + "learning_rate": 0.0003418103296373096, + "loss": 3.4139351844787598, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.022404303774237633, + "learning_rate": 0.00034154601335542964, + "loss": 3.4200823307037354, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.021705076098442078, + "learning_rate": 0.00034128166601772304, + "loss": 3.4092609882354736, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.02273249812424183, + "learning_rate": 0.0003410172878371695, + "loss": 3.3931846618652344, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.022552402690052986, + "learning_rate": 0.00034075287902677394, + "loss": 3.396566867828369, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.023219924420118332, + "learning_rate": 0.0003404884397995655, + "loss": 3.369025707244873, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.022560900077223778, + "learning_rate": 0.00034022397036859837, + "loss": 3.4248225688934326, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.023370858281850815, + "learning_rate": 0.0003399594709469506, + "loss": 3.431628942489624, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.02469109185039997, + "learning_rate": 0.00033969494174772465, + "loss": 3.406843662261963, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.02417670376598835, + "learning_rate": 0.0003394303829840469, + "loss": 3.445046901702881, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.0237257182598114, + "learning_rate": 0.00033916579486906764, + "loss": 3.4490597248077393, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.02295520156621933, + "learning_rate": 0.0003389011776159607, + "loss": 3.4009876251220703, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.023158475756645203, + "learning_rate": 0.0003386365314379233, + "loss": 3.4383902549743652, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.023618478327989578, + "learning_rate": 0.0003383718565481763, + "loss": 3.421168804168701, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.022055329754948616, + "learning_rate": 0.0003381071531599633, + "loss": 3.382946729660034, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.021441346034407616, + "learning_rate": 0.00033784242148655115, + "loss": 3.4464712142944336, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.023031961172819138, + "learning_rate": 0.00033757766174122934, + "loss": 3.42844295501709, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.022832952439785004, + "learning_rate": 0.00033731287413731005, + "loss": 3.398231029510498, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.022094765678048134, + "learning_rate": 0.0003370480588881278, + "loss": 3.4270074367523193, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.0218376275151968, + "learning_rate": 0.0003367832162070395, + "loss": 3.4681742191314697, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.02364739589393139, + "learning_rate": 0.000336518346307424, + "loss": 3.431920051574707, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.023402225226163864, + "learning_rate": 0.0003362534494026824, + "loss": 3.4189023971557617, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.022775836288928986, + "learning_rate": 0.0003359885257062372, + "loss": 3.409396171569824, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.02445383556187153, + "learning_rate": 0.0003357235754315328, + "loss": 3.41017746925354, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.066352367401123, + "eval_runtime": 58.1658, + "eval_samples_per_second": 41.983, + "eval_steps_per_second": 1.324, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.02710847556591034, + "learning_rate": 0.0003354585987920345, + "loss": 3.3155131340026855, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.03428799286484718, + "learning_rate": 0.0003351935960012296, + "loss": 3.327408790588379, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.03434379771351814, + "learning_rate": 0.0003349285672726259, + "loss": 3.3282251358032227, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.03405233472585678, + "learning_rate": 0.0003346635128197522, + "loss": 3.2932448387145996, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.031194346025586128, + "learning_rate": 0.0003343984328561581, + "loss": 3.336615800857544, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.031816478818655014, + "learning_rate": 0.00033413332759541376, + "loss": 3.3375275135040283, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.03413299471139908, + "learning_rate": 0.0003338681972511098, + "loss": 3.3710694313049316, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.034200165420770645, + "learning_rate": 0.0003336030420368568, + "loss": 3.3248214721679688, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.030694160610437393, + "learning_rate": 0.0003333378621662857, + "loss": 3.3297901153564453, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.031172804534435272, + "learning_rate": 0.00033307265785304684, + "loss": 3.287217140197754, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.031914614140987396, + "learning_rate": 0.00033280742931081076, + "loss": 3.3218531608581543, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.0332055389881134, + "learning_rate": 0.00033254217675326737, + "loss": 3.3454785346984863, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.034956980496644974, + "learning_rate": 0.0003322769003941257, + "loss": 3.3193702697753906, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.03618055582046509, + "learning_rate": 0.00033201160044711423, + "loss": 3.3661017417907715, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.030322827398777008, + "learning_rate": 0.0003317462771259802, + "loss": 3.353989601135254, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.029768960550427437, + "learning_rate": 0.00033148093064449006, + "loss": 3.3254497051239014, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.028036782518029213, + "learning_rate": 0.0003312155612164284, + "loss": 3.3276920318603516, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.028774483129382133, + "learning_rate": 0.00033095016905559883, + "loss": 3.3458826541900635, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.030403515323996544, + "learning_rate": 0.0003306847543758227, + "loss": 3.3332793712615967, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.028468478471040726, + "learning_rate": 0.0003304193173909401, + "loss": 3.3167755603790283, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.029126476496458054, + "learning_rate": 0.00033015385831480873, + "loss": 3.299426794052124, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.030746638774871826, + "learning_rate": 0.0003298883773613043, + "loss": 3.3581809997558594, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.030241351574659348, + "learning_rate": 0.0003296228747443197, + "loss": 3.3415942192077637, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.029754415154457092, + "learning_rate": 0.0003293573506777659, + "loss": 3.3525662422180176, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.026896746829152107, + "learning_rate": 0.00032909180537557076, + "loss": 3.321808099746704, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.026361066848039627, + "learning_rate": 0.00032882623905167917, + "loss": 3.329258441925049, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.027747981250286102, + "learning_rate": 0.00032856065192005335, + "loss": 3.3774523735046387, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.026270100846886635, + "learning_rate": 0.00032829504419467194, + "loss": 3.3419976234436035, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.02731695957481861, + "learning_rate": 0.0003280294160895303, + "loss": 3.3068814277648926, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.026583299040794373, + "learning_rate": 0.0003277637678186402, + "loss": 3.303804874420166, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.026615949347615242, + "learning_rate": 0.00032749809959602973, + "loss": 3.3812599182128906, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.027221621945500374, + "learning_rate": 0.0003272324116357428, + "loss": 3.327211618423462, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.025025667622685432, + "learning_rate": 0.0003269667041518395, + "loss": 3.339163303375244, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.02593836560845375, + "learning_rate": 0.0003267009773583956, + "loss": 3.3517372608184814, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.023697076365351677, + "learning_rate": 0.00032643523146950234, + "loss": 3.379819869995117, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.02425902709364891, + "learning_rate": 0.0003261694666992664, + "loss": 3.390653610229492, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.025133250281214714, + "learning_rate": 0.0003259036832618096, + "loss": 3.364495277404785, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.025542905554175377, + "learning_rate": 0.0003256378813712688, + "loss": 3.3584814071655273, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.026041125878691673, + "learning_rate": 0.000325372061241796, + "loss": 3.36777400970459, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.02391037344932556, + "learning_rate": 0.00032510622308755746, + "loss": 3.3809385299682617, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.02379394695162773, + "learning_rate": 0.00032484036712273424, + "loss": 3.354257583618164, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.0234722550958395, + "learning_rate": 0.0003245744935615219, + "loss": 3.3736205101013184, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.024123650044202805, + "learning_rate": 0.0003243086026181296, + "loss": 3.3589959144592285, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.024013696238398552, + "learning_rate": 0.00032404269450678116, + "loss": 3.353874444961548, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.022617843002080917, + "learning_rate": 0.00032377676944171375, + "loss": 3.364056348800659, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.022730547934770584, + "learning_rate": 0.00032351082763717857, + "loss": 3.378067970275879, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.022257911041378975, + "learning_rate": 0.0003232448693074399, + "loss": 3.379342555999756, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.023247094824910164, + "learning_rate": 0.00032297889466677575, + "loss": 3.3537116050720215, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.022764021530747414, + "learning_rate": 0.000322712903929477, + "loss": 3.3969063758850098, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.02260403335094452, + "learning_rate": 0.0003224468973098477, + "loss": 3.3727574348449707, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.021772351115942, + "learning_rate": 0.0003221808750222044, + "loss": 3.335434675216675, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.02238459885120392, + "learning_rate": 0.0003219148372808766, + "loss": 3.3440277576446533, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.0230540968477726, + "learning_rate": 0.00032164878430020606, + "loss": 3.3997583389282227, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.0220821350812912, + "learning_rate": 0.00032138271629454684, + "loss": 3.377815008163452, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.021497417241334915, + "learning_rate": 0.00032111663347826505, + "loss": 3.3696370124816895, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.02163885347545147, + "learning_rate": 0.00032085053606573896, + "loss": 3.3489794731140137, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.022394269704818726, + "learning_rate": 0.0003205844242713584, + "loss": 3.3792812824249268, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.022390708327293396, + "learning_rate": 0.0003203182983095248, + "loss": 3.3733272552490234, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.023712158203125, + "learning_rate": 0.0003200521583946511, + "loss": 3.3716251850128174, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.02398860827088356, + "learning_rate": 0.0003197860047411613, + "loss": 3.381779670715332, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.02309161052107811, + "learning_rate": 0.00031951983756349066, + "loss": 3.3923277854919434, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.021400388330221176, + "learning_rate": 0.00031925365707608536, + "loss": 3.407397747039795, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.022534485906362534, + "learning_rate": 0.00031898746349340204, + "loss": 3.3740391731262207, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.023533202707767487, + "learning_rate": 0.0003187212570299082, + "loss": 3.3877944946289062, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.023398835211992264, + "learning_rate": 0.0003184550379000815, + "loss": 3.3936471939086914, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.020927201956510544, + "learning_rate": 0.00031818880631841005, + "loss": 3.339545249938965, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.021453995257616043, + "learning_rate": 0.0003179225624993915, + "loss": 3.3647522926330566, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.022847114130854607, + "learning_rate": 0.0003176563066575341, + "loss": 3.405144691467285, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.022766996175050735, + "learning_rate": 0.0003173900390073549, + "loss": 3.392810344696045, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.022103948518633842, + "learning_rate": 0.0003171237597633813, + "loss": 3.3973004817962646, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.022282181307673454, + "learning_rate": 0.0003168574691401496, + "loss": 3.409998893737793, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.025435522198677063, + "learning_rate": 0.0003165911673522053, + "loss": 3.423318386077881, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.078423500061035, + "eval_runtime": 57.3189, + "eval_samples_per_second": 42.604, + "eval_steps_per_second": 1.343, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.02624760940670967, + "learning_rate": 0.00031632485461410294, + "loss": 3.261288642883301, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.03552987053990364, + "learning_rate": 0.00031605853114040583, + "loss": 3.2682106494903564, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.040380869060754776, + "learning_rate": 0.00031579219714568613, + "loss": 3.2813782691955566, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.03861293941736221, + "learning_rate": 0.0003155258528445242, + "loss": 3.258193016052246, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.04091731086373329, + "learning_rate": 0.0003152594984515089, + "loss": 3.3052220344543457, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.043269723653793335, + "learning_rate": 0.00031499313418123704, + "loss": 3.2820217609405518, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.04580320790410042, + "learning_rate": 0.0003147267602483138, + "loss": 3.312005043029785, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.047336459159851074, + "learning_rate": 0.00031446037686735153, + "loss": 3.3217713832855225, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.04587360844016075, + "learning_rate": 0.0003141939842529706, + "loss": 3.2869949340820312, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.04002532362937927, + "learning_rate": 0.0003139275826197987, + "loss": 3.2879223823547363, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.0368693508207798, + "learning_rate": 0.00031366117218247094, + "loss": 3.2983577251434326, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.037317775189876556, + "learning_rate": 0.00031339475315562916, + "loss": 3.3062524795532227, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.03655700013041496, + "learning_rate": 0.00031312832575392267, + "loss": 3.31072735786438, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.033466216176748276, + "learning_rate": 0.0003128618901920071, + "loss": 3.2954514026641846, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.030785970389842987, + "learning_rate": 0.0003125954466845446, + "loss": 3.27952241897583, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.029669342562556267, + "learning_rate": 0.00031232899544620404, + "loss": 3.2613229751586914, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.033505845814943314, + "learning_rate": 0.00031206253669166045, + "loss": 3.2987258434295654, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.032395437359809875, + "learning_rate": 0.0003117960706355947, + "loss": 3.296708583831787, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.031322430819272995, + "learning_rate": 0.0003115295974926936, + "loss": 3.3160252571105957, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.029345186427235603, + "learning_rate": 0.00031126311747765005, + "loss": 3.307143211364746, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.026715675368905067, + "learning_rate": 0.00031099663080516186, + "loss": 3.286012649536133, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.028592750430107117, + "learning_rate": 0.0003107301376899327, + "loss": 3.2528645992279053, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.028861764818429947, + "learning_rate": 0.00031046363834667123, + "loss": 3.2795748710632324, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.02609528973698616, + "learning_rate": 0.00031019713299009123, + "loss": 3.326317071914673, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.027718454599380493, + "learning_rate": 0.000309930621834911, + "loss": 3.268110513687134, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.026382867246866226, + "learning_rate": 0.0003096641050958541, + "loss": 3.294499397277832, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.027083583176136017, + "learning_rate": 0.00030939758298764795, + "loss": 3.3202130794525146, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.02516467310488224, + "learning_rate": 0.00030913105572502483, + "loss": 3.3485021591186523, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.024148069322109222, + "learning_rate": 0.00030886452352272095, + "loss": 3.3099541664123535, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.024552857503294945, + "learning_rate": 0.00030859798659547624, + "loss": 3.2885942459106445, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.02551114559173584, + "learning_rate": 0.0003083314451580349, + "loss": 3.3232040405273438, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.025423647835850716, + "learning_rate": 0.0003080648994251445, + "loss": 3.3097832202911377, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.023211417719721794, + "learning_rate": 0.0003077983496115561, + "loss": 3.3075966835021973, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.024623898789286613, + "learning_rate": 0.00030753179593202406, + "loss": 3.2917377948760986, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.02677663415670395, + "learning_rate": 0.0003072652386013059, + "loss": 3.356265068054199, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.02673642337322235, + "learning_rate": 0.00030699867783416207, + "loss": 3.314903736114502, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.024466225877404213, + "learning_rate": 0.0003067321138453557, + "loss": 3.30912446975708, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.025370748713612556, + "learning_rate": 0.0003064655468496527, + "loss": 3.363640069961548, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.026536114513874054, + "learning_rate": 0.0003061989770618211, + "loss": 3.304851531982422, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.026536893099546432, + "learning_rate": 0.00030593240469663154, + "loss": 3.342902898788452, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.02645307406783104, + "learning_rate": 0.0003056658299688563, + "loss": 3.3164432048797607, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.02523060515522957, + "learning_rate": 0.00030539925309327017, + "loss": 3.34376859664917, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.02590431459248066, + "learning_rate": 0.0003051326742846491, + "loss": 3.3479061126708984, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.0230655986815691, + "learning_rate": 0.000304866093757771, + "loss": 3.2890982627868652, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.02403893694281578, + "learning_rate": 0.00030459951172741474, + "loss": 3.345456600189209, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.02238907851278782, + "learning_rate": 0.0003043329284083609, + "loss": 3.3471767902374268, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.021377569064497948, + "learning_rate": 0.00030406634401539053, + "loss": 3.319232225418091, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.02272353693842888, + "learning_rate": 0.00030379975876328614, + "loss": 3.316892623901367, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.02413824200630188, + "learning_rate": 0.00030353317286683053, + "loss": 3.3450968265533447, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.02334682270884514, + "learning_rate": 0.0003032665865408073, + "loss": 3.3640642166137695, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.021966520696878433, + "learning_rate": 0.000303, + "loss": 3.358555316925049, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.023984894156455994, + "learning_rate": 0.0003027334134591926, + "loss": 3.358398914337158, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.023520268499851227, + "learning_rate": 0.00030246682713316945, + "loss": 3.3224587440490723, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.0231285709887743, + "learning_rate": 0.0003022002412367138, + "loss": 3.3671562671661377, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.022818170487880707, + "learning_rate": 0.00030193365598460946, + "loss": 3.361027717590332, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.02280694991350174, + "learning_rate": 0.0003016670715916391, + "loss": 3.2923383712768555, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.021362345665693283, + "learning_rate": 0.00030140048827258524, + "loss": 3.3346734046936035, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.02372218668460846, + "learning_rate": 0.0003011339062422289, + "loss": 3.354748249053955, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.022949764505028725, + "learning_rate": 0.00030086732571535076, + "loss": 3.3857789039611816, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.021954255178570747, + "learning_rate": 0.0003006007469067297, + "loss": 3.343330144882202, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.022438595071434975, + "learning_rate": 0.0003003341700311436, + "loss": 3.328618049621582, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.021969925612211227, + "learning_rate": 0.0003000675953033685, + "loss": 3.355971574783325, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.022706981748342514, + "learning_rate": 0.0002998010229381789, + "loss": 3.3529577255249023, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.024083172902464867, + "learning_rate": 0.0002995344531503474, + "loss": 3.348170757293701, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.023259354755282402, + "learning_rate": 0.0002992678861546442, + "loss": 3.3474369049072266, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.02218157798051834, + "learning_rate": 0.00029900132216583786, + "loss": 3.369053840637207, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.02305331453680992, + "learning_rate": 0.0002987347613986941, + "loss": 3.3622641563415527, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.022796178236603737, + "learning_rate": 0.0002984682040679759, + "loss": 3.351510524749756, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.021299205720424652, + "learning_rate": 0.00029820165038844395, + "loss": 3.339406728744507, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.022786924615502357, + "learning_rate": 0.00029793510057485556, + "loss": 3.3182196617126465, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.021729128435254097, + "learning_rate": 0.00029766855484196513, + "loss": 3.383904218673706, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.024604283273220062, + "learning_rate": 0.0002974020134045238, + "loss": 3.2912864685058594, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.0894880294799805, + "eval_runtime": 58.3967, + "eval_samples_per_second": 41.817, + "eval_steps_per_second": 1.319, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.02673362009227276, + "learning_rate": 0.00029713547647727904, + "loss": 3.270986795425415, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.04036599025130272, + "learning_rate": 0.00029686894427497515, + "loss": 3.2358999252319336, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.047250088304281235, + "learning_rate": 0.0002966024170123519, + "loss": 3.2706189155578613, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.04642819985747337, + "learning_rate": 0.0002963358949041459, + "loss": 3.236340284347534, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.039405327290296555, + "learning_rate": 0.00029606937816508897, + "loss": 3.203136682510376, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.038748081773519516, + "learning_rate": 0.00029580286700990887, + "loss": 3.2636070251464844, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.039233580231666565, + "learning_rate": 0.00029553636165332876, + "loss": 3.2751896381378174, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.03429320082068443, + "learning_rate": 0.0002952698623100673, + "loss": 3.2765679359436035, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.03267038241028786, + "learning_rate": 0.0002950033691948382, + "loss": 3.2387681007385254, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.03349406272172928, + "learning_rate": 0.00029473688252234994, + "loss": 3.282162666320801, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.032002877444028854, + "learning_rate": 0.0002944704025073064, + "loss": 3.246623992919922, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.031509071588516235, + "learning_rate": 0.00029420392936440525, + "loss": 3.2374932765960693, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.02958357334136963, + "learning_rate": 0.00029393746330833953, + "loss": 3.24362850189209, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.03248635679483414, + "learning_rate": 0.0002936710045537959, + "loss": 3.237210273742676, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.03230655938386917, + "learning_rate": 0.0002934045533154554, + "loss": 3.251129150390625, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.029688173905014992, + "learning_rate": 0.00029313810980799296, + "loss": 3.2141575813293457, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.027773940935730934, + "learning_rate": 0.00029287167424607726, + "loss": 3.264249086380005, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.02687031403183937, + "learning_rate": 0.00029260524684437077, + "loss": 3.2513632774353027, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.02503623627126217, + "learning_rate": 0.00029233882781752905, + "loss": 3.217961072921753, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.027071228250861168, + "learning_rate": 0.00029207241738020136, + "loss": 3.2784645557403564, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.02600773423910141, + "learning_rate": 0.0002918060157470294, + "loss": 3.2641382217407227, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.026998162269592285, + "learning_rate": 0.00029153962313264856, + "loss": 3.2646782398223877, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.025579402223229408, + "learning_rate": 0.0002912732397516862, + "loss": 3.2700581550598145, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.02645333670079708, + "learning_rate": 0.00029100686581876283, + "loss": 3.2545266151428223, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.02717350795865059, + "learning_rate": 0.0002907405015484911, + "loss": 3.2795705795288086, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.02675914391875267, + "learning_rate": 0.0002904741471554758, + "loss": 3.2936904430389404, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.02751768007874489, + "learning_rate": 0.0002902078028543139, + "loss": 3.2329440116882324, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.025446882471442223, + "learning_rate": 0.00028994146885959416, + "loss": 3.274294376373291, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.02365768514573574, + "learning_rate": 0.00028967514538589715, + "loss": 3.2585229873657227, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.0266884732991457, + "learning_rate": 0.0002894088326477947, + "loss": 3.2872962951660156, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.02572469227015972, + "learning_rate": 0.0002891425308598503, + "loss": 3.253828287124634, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.02540273219347, + "learning_rate": 0.00028887624023661866, + "loss": 3.303926944732666, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.024694625288248062, + "learning_rate": 0.00028860996099264505, + "loss": 3.264145851135254, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.027233170345425606, + "learning_rate": 0.000288343693342466, + "loss": 3.286924362182617, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02635285072028637, + "learning_rate": 0.00028807743750060847, + "loss": 3.2930331230163574, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.025157805532217026, + "learning_rate": 0.00028781119368159004, + "loss": 3.2906906604766846, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.024729596450924873, + "learning_rate": 0.00028754496209991843, + "loss": 3.277128219604492, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.023809000849723816, + "learning_rate": 0.0002872787429700917, + "loss": 3.2957217693328857, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02474421076476574, + "learning_rate": 0.0002870125365065979, + "loss": 3.27791690826416, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.02439294196665287, + "learning_rate": 0.00028674634292391463, + "loss": 3.2912981510162354, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.023601334542036057, + "learning_rate": 0.0002864801624365093, + "loss": 3.317446231842041, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.0253264419734478, + "learning_rate": 0.0002862139952588387, + "loss": 3.2742302417755127, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.02389134280383587, + "learning_rate": 0.000285947841605349, + "loss": 3.281032085418701, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.02451813407242298, + "learning_rate": 0.00028568170169047523, + "loss": 3.3339295387268066, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.024324845522642136, + "learning_rate": 0.00028541557572864154, + "loss": 3.290402889251709, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.024939008057117462, + "learning_rate": 0.000285149463934261, + "loss": 3.2933621406555176, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.024450505152344704, + "learning_rate": 0.0002848833665217349, + "loss": 3.3055014610290527, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.02262270078063011, + "learning_rate": 0.0002846172837054532, + "loss": 3.2964205741882324, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.02311946265399456, + "learning_rate": 0.0002843512156997939, + "loss": 3.3102402687072754, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.024362705647945404, + "learning_rate": 0.0002840851627191234, + "loss": 3.3020925521850586, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.024785032495856285, + "learning_rate": 0.0002838191249777955, + "loss": 3.287290096282959, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.02358846180140972, + "learning_rate": 0.00028355310269015225, + "loss": 3.334261894226074, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.024330584332346916, + "learning_rate": 0.00028328709607052297, + "loss": 3.2799081802368164, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.023546984419226646, + "learning_rate": 0.0002830211053332242, + "loss": 3.3228847980499268, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.023060832172632217, + "learning_rate": 0.0002827551306925601, + "loss": 3.325793981552124, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.02388455718755722, + "learning_rate": 0.00028248917236282147, + "loss": 3.2892062664031982, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.023740701377391815, + "learning_rate": 0.0002822232305582863, + "loss": 3.2969369888305664, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.02458217367529869, + "learning_rate": 0.0002819573054932188, + "loss": 3.315610408782959, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.023173579946160316, + "learning_rate": 0.0002816913973818703, + "loss": 3.2561936378479004, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.024137936532497406, + "learning_rate": 0.0002814255064384781, + "loss": 3.2839202880859375, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.022436272352933884, + "learning_rate": 0.00028115963287726563, + "loss": 3.3057351112365723, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.023997878655791283, + "learning_rate": 0.0002808937769124425, + "loss": 3.3583154678344727, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.024119896814227104, + "learning_rate": 0.000280627938758204, + "loss": 3.322892427444458, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.022283365949988365, + "learning_rate": 0.00028036211862873117, + "loss": 3.3145735263824463, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.023433363065123558, + "learning_rate": 0.0002800963167381904, + "loss": 3.3564529418945312, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.02224746160209179, + "learning_rate": 0.00027983053330073355, + "loss": 3.2985916137695312, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.023786570876836777, + "learning_rate": 0.00027956476853049765, + "loss": 3.302539587020874, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.025200961157679558, + "learning_rate": 0.00027929902264160435, + "loss": 3.309976577758789, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.023849809542298317, + "learning_rate": 0.00027903329584816046, + "loss": 3.343660354614258, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.023311031982302666, + "learning_rate": 0.0002787675883642572, + "loss": 3.2978909015655518, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.023174891248345375, + "learning_rate": 0.00027850190040397036, + "loss": 3.326188802719116, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.02681034244596958, + "learning_rate": 0.00027823623218135976, + "loss": 3.3422017097473145, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.109009265899658, + "eval_runtime": 58.0506, + "eval_samples_per_second": 42.067, + "eval_steps_per_second": 1.326, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.028057295829057693, + "learning_rate": 0.0002779705839104697, + "loss": 3.1825037002563477, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.034786805510520935, + "learning_rate": 0.0002777049558053281, + "loss": 3.200389862060547, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.03751441463828087, + "learning_rate": 0.0002774393480799466, + "loss": 3.2047512531280518, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.035590436309576035, + "learning_rate": 0.00027717376094832076, + "loss": 3.1735124588012695, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.03426426649093628, + "learning_rate": 0.0002769081946244293, + "loss": 3.199483871459961, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.037030890583992004, + "learning_rate": 0.00027664264932223407, + "loss": 3.213322639465332, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.032468877732753754, + "learning_rate": 0.00027637712525568024, + "loss": 3.179354190826416, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.027202758938074112, + "learning_rate": 0.0002761116226386958, + "loss": 3.216036319732666, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.031184563413262367, + "learning_rate": 0.00027584614168519125, + "loss": 3.208205223083496, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.029414812102913857, + "learning_rate": 0.00027558068260905976, + "loss": 3.180370569229126, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.02679901383817196, + "learning_rate": 0.00027531524562417734, + "loss": 3.210325002670288, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.02863631211221218, + "learning_rate": 0.00027504983094440126, + "loss": 3.1601076126098633, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.029135366901755333, + "learning_rate": 0.0002747844387835716, + "loss": 3.2262887954711914, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.030535712838172913, + "learning_rate": 0.0002745190693555099, + "loss": 3.23160719871521, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.030563877895474434, + "learning_rate": 0.0002742537228740197, + "loss": 3.255211353302002, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.02906505949795246, + "learning_rate": 0.00027398839955288575, + "loss": 3.2006797790527344, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.030871668830513954, + "learning_rate": 0.00027372309960587424, + "loss": 3.2442867755889893, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.02913421392440796, + "learning_rate": 0.0002734578232467327, + "loss": 3.218442440032959, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.030087953433394432, + "learning_rate": 0.0002731925706891891, + "loss": 3.2164535522460938, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.034147944301366806, + "learning_rate": 0.0002729273421469532, + "loss": 3.218660831451416, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.03593966364860535, + "learning_rate": 0.00027266213783371434, + "loss": 3.2240614891052246, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.032090652734041214, + "learning_rate": 0.00027239695796314327, + "loss": 3.2309441566467285, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.02776424027979374, + "learning_rate": 0.0002721318027488902, + "loss": 3.2702016830444336, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.027174271643161774, + "learning_rate": 0.00027186667240458617, + "loss": 3.19810152053833, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.028479408472776413, + "learning_rate": 0.0002716015671438419, + "loss": 3.2178421020507812, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.026820644736289978, + "learning_rate": 0.0002713364871802478, + "loss": 3.2634482383728027, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.026427917182445526, + "learning_rate": 0.0002710714327273741, + "loss": 3.2452914714813232, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.02679610252380371, + "learning_rate": 0.0002708064039987704, + "loss": 3.243825912475586, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.026115626096725464, + "learning_rate": 0.00027054140120796546, + "loss": 3.218787670135498, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.026686571538448334, + "learning_rate": 0.0002702764245684672, + "loss": 3.2785303592681885, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.02685859426856041, + "learning_rate": 0.0002700114742937627, + "loss": 3.2572226524353027, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.026968171820044518, + "learning_rate": 0.00026974655059731754, + "loss": 3.2625553607940674, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.025397246703505516, + "learning_rate": 0.0002694816536925759, + "loss": 3.262303352355957, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.025660471990704536, + "learning_rate": 0.00026921678379296057, + "loss": 3.269401788711548, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.02713891491293907, + "learning_rate": 0.0002689519411118722, + "loss": 3.2416529655456543, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.027591092512011528, + "learning_rate": 0.00026868712586269, + "loss": 3.2814316749572754, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.024462848901748657, + "learning_rate": 0.00026842233825877064, + "loss": 3.249725103378296, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.024868478998541832, + "learning_rate": 0.0002681575785134487, + "loss": 3.231574535369873, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.027015401050448418, + "learning_rate": 0.0002678928468400366, + "loss": 3.282627582550049, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.02678140625357628, + "learning_rate": 0.00026762814345182367, + "loss": 3.2208900451660156, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.026830527931451797, + "learning_rate": 0.0002673634685620767, + "loss": 3.260575294494629, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.02430890128016472, + "learning_rate": 0.0002670988223840393, + "loss": 3.2522997856140137, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.027047613635659218, + "learning_rate": 0.0002668342051309324, + "loss": 3.2721078395843506, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.026861369609832764, + "learning_rate": 0.00026656961701595306, + "loss": 3.2235946655273438, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.026650549843907356, + "learning_rate": 0.0002663050582522753, + "loss": 3.302743434906006, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.02551819011569023, + "learning_rate": 0.00026604052905304946, + "loss": 3.265763282775879, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.024843450635671616, + "learning_rate": 0.00026577602963140156, + "loss": 3.272397518157959, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.024423206225037575, + "learning_rate": 0.0002655115602004345, + "loss": 3.253495454788208, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.024180661886930466, + "learning_rate": 0.00026524712097322604, + "loss": 3.260438919067383, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.023554982617497444, + "learning_rate": 0.0002649827121628305, + "loss": 3.2342851161956787, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.023470086976885796, + "learning_rate": 0.00026471833398227694, + "loss": 3.251622438430786, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.024405231699347496, + "learning_rate": 0.0002644539866445703, + "loss": 3.2588815689086914, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.024147814139723778, + "learning_rate": 0.00026418967036269045, + "loss": 3.2723565101623535, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.023217910900712013, + "learning_rate": 0.0002639253853495921, + "loss": 3.2555792331695557, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.024921158328652382, + "learning_rate": 0.00026366113181820506, + "loss": 3.26884126663208, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.02455919235944748, + "learning_rate": 0.00026339690998143346, + "loss": 3.2674927711486816, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.024725932627916336, + "learning_rate": 0.00026313272005215637, + "loss": 3.2511110305786133, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.025769414380192757, + "learning_rate": 0.00026286856224322645, + "loss": 3.290651321411133, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.025928806513547897, + "learning_rate": 0.00026260443676747114, + "loss": 3.302396297454834, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.02499072253704071, + "learning_rate": 0.00026234034383769157, + "loss": 3.2813124656677246, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.025431277230381966, + "learning_rate": 0.0002620762836666625, + "loss": 3.234684705734253, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.026871860027313232, + "learning_rate": 0.00026181225646713257, + "loss": 3.2800331115722656, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.02578776702284813, + "learning_rate": 0.0002615482624518236, + "loss": 3.3117146492004395, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.026318378746509552, + "learning_rate": 0.0002612843018334309, + "loss": 3.257615089416504, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.025542527437210083, + "learning_rate": 0.00026102037482462256, + "loss": 3.2969868183135986, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.025723429396748543, + "learning_rate": 0.0002607564816380403, + "loss": 3.290259838104248, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.024588774889707565, + "learning_rate": 0.0002604926224862975, + "loss": 3.284090280532837, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.024619122967123985, + "learning_rate": 0.0002602287975819811, + "loss": 3.284879207611084, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.024441618472337723, + "learning_rate": 0.00025996500713765, + "loss": 3.283104658126831, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.02402760274708271, + "learning_rate": 0.0002597012513658352, + "loss": 3.3013617992401123, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.025028521195054054, + "learning_rate": 0.00025943753047904026, + "loss": 3.2892568111419678, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.029445001855492592, + "learning_rate": 0.0002591738446897401, + "loss": 3.2474446296691895, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.128176689147949, + "eval_runtime": 58.8588, + "eval_samples_per_second": 41.489, + "eval_steps_per_second": 1.308, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.0298826452344656, + "learning_rate": 0.00025891019421038177, + "loss": 3.138103485107422, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.0387548953294754, + "learning_rate": 0.00025864657925338345, + "loss": 3.161492347717285, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.03816070780158043, + "learning_rate": 0.00025838300003113514, + "loss": 3.1739673614501953, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.03409738838672638, + "learning_rate": 0.0002581194567559979, + "loss": 3.182478189468384, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.033135563135147095, + "learning_rate": 0.00025785594964030365, + "loss": 3.1720199584960938, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.03203624114394188, + "learning_rate": 0.0002575924788963554, + "loss": 3.1699862480163574, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.031237754970788956, + "learning_rate": 0.0002573290447364268, + "loss": 3.179363250732422, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.03342146426439285, + "learning_rate": 0.0002570656473727619, + "loss": 3.1822280883789062, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.03561089560389519, + "learning_rate": 0.0002568022870175752, + "loss": 3.1767094135284424, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.034710463136434555, + "learning_rate": 0.00025653896388305134, + "loss": 3.218817949295044, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.030870331451296806, + "learning_rate": 0.0002562756781813453, + "loss": 3.168602705001831, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.028004391118884087, + "learning_rate": 0.00025601243012458126, + "loss": 3.224155902862549, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.030953701585531235, + "learning_rate": 0.00025574921992485366, + "loss": 3.1789557933807373, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.030289432033896446, + "learning_rate": 0.000255486047794226, + "loss": 3.1890387535095215, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.030684132128953934, + "learning_rate": 0.0002552229139447316, + "loss": 3.195427417755127, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.03022196516394615, + "learning_rate": 0.00025495981858837235, + "loss": 3.1984965801239014, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.026935536414384842, + "learning_rate": 0.0002546967619371195, + "loss": 3.2014272212982178, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.02702922560274601, + "learning_rate": 0.0002544337442029133, + "loss": 3.205538034439087, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.026682600378990173, + "learning_rate": 0.00025417076559766194, + "loss": 3.1954591274261475, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.0273384228348732, + "learning_rate": 0.000253907826333243, + "loss": 3.1782338619232178, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.028587577864527702, + "learning_rate": 0.00025364492662150145, + "loss": 3.199565887451172, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.028848132118582726, + "learning_rate": 0.0002533820666742514, + "loss": 3.181786298751831, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.027535341680049896, + "learning_rate": 0.00025311924670327364, + "loss": 3.2013628482818604, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.02756025455892086, + "learning_rate": 0.00025285646692031804, + "loss": 3.1749587059020996, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.027358995750546455, + "learning_rate": 0.0002525937275371013, + "loss": 3.161881923675537, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.025918981060385704, + "learning_rate": 0.00025233102876530785, + "loss": 3.225224256515503, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.026479976251721382, + "learning_rate": 0.0002520683708165894, + "loss": 3.2129316329956055, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.02685079537332058, + "learning_rate": 0.0002518057539025646, + "loss": 3.1651554107666016, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.027378706261515617, + "learning_rate": 0.00025154317823481917, + "loss": 3.193108558654785, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.027240974828600883, + "learning_rate": 0.0002512806440249056, + "loss": 3.190605640411377, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.028065402060747147, + "learning_rate": 0.00025101815148434296, + "loss": 3.2068259716033936, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.028329014778137207, + "learning_rate": 0.00025075570082461684, + "loss": 3.2316060066223145, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.028920065611600876, + "learning_rate": 0.0002504932922571789, + "loss": 3.1933932304382324, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.028299354016780853, + "learning_rate": 0.00025023092599344706, + "loss": 3.192711353302002, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.025996020063757896, + "learning_rate": 0.0002499686022448049, + "loss": 3.243171453475952, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.0264595914632082, + "learning_rate": 0.00024970632122260216, + "loss": 3.199758529663086, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.02704528719186783, + "learning_rate": 0.0002494440831381538, + "loss": 3.2390713691711426, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.026571933180093765, + "learning_rate": 0.00024918188820274016, + "loss": 3.219440460205078, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.025562983006238937, + "learning_rate": 0.0002489197366276073, + "loss": 3.2538228034973145, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.024683184921741486, + "learning_rate": 0.0002486576286239657, + "loss": 3.2372591495513916, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.025656528770923615, + "learning_rate": 0.0002483955644029913, + "loss": 3.250558853149414, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.024807540699839592, + "learning_rate": 0.0002481335441758243, + "loss": 3.1856563091278076, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.026149194687604904, + "learning_rate": 0.0002478715681535699, + "loss": 3.172186851501465, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.02617705799639225, + "learning_rate": 0.00024760963654729717, + "loss": 3.250445604324341, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.026542512699961662, + "learning_rate": 0.0002473477495680397, + "loss": 3.220273494720459, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.024963509291410446, + "learning_rate": 0.00024708590742679546, + "loss": 3.2330946922302246, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.027144264429807663, + "learning_rate": 0.00024682411033452546, + "loss": 3.1860971450805664, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.02665683813393116, + "learning_rate": 0.0002465623585021553, + "loss": 3.257209062576294, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.024872658774256706, + "learning_rate": 0.0002463006521405733, + "loss": 3.2266182899475098, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.025235185399651527, + "learning_rate": 0.0002460389914606319, + "loss": 3.2134504318237305, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.025029495358467102, + "learning_rate": 0.00024577737667314615, + "loss": 3.2150087356567383, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.025630896911025047, + "learning_rate": 0.00024551580798889446, + "loss": 3.282879114151001, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.02722341939806938, + "learning_rate": 0.00024525428561861804, + "loss": 3.2366206645965576, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.02522372268140316, + "learning_rate": 0.0002449928097730207, + "loss": 3.2840523719787598, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.02494790591299534, + "learning_rate": 0.00024473138066276876, + "loss": 3.2405319213867188, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.026662737131118774, + "learning_rate": 0.0002444699984984909, + "loss": 3.239692211151123, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.026082608848810196, + "learning_rate": 0.0002442086634907782, + "loss": 3.2156195640563965, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.02423974685370922, + "learning_rate": 0.00024394737585018328, + "loss": 3.1863760948181152, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.025226688012480736, + "learning_rate": 0.00024368613578722113, + "loss": 3.20874285697937, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.02428225241601467, + "learning_rate": 0.00024342494351236799, + "loss": 3.2209601402282715, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.023819707334041595, + "learning_rate": 0.0002431637992360618, + "loss": 3.2564258575439453, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.025661710649728775, + "learning_rate": 0.00024290270316870184, + "loss": 3.244133949279785, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.02423049882054329, + "learning_rate": 0.0002426416555206484, + "loss": 3.2376630306243896, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.02382473088800907, + "learning_rate": 0.000242380656502223, + "loss": 3.2951290607452393, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.02488875202834606, + "learning_rate": 0.00024211970632370756, + "loss": 3.2848310470581055, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.02575223706662655, + "learning_rate": 0.0002418588051953453, + "loss": 3.2574503421783447, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.02462848648428917, + "learning_rate": 0.0002415979533273392, + "loss": 3.2464442253112793, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.024912388995289803, + "learning_rate": 0.0002413371509298531, + "loss": 3.2459192276000977, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.025020884349942207, + "learning_rate": 0.00024107639821301078, + "loss": 3.213622808456421, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.026330886408686638, + "learning_rate": 0.00024081569538689587, + "loss": 3.2305431365966797, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.02708602324128151, + "learning_rate": 0.0002405550426615521, + "loss": 3.2487688064575195, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.030908891931176186, + "learning_rate": 0.00024029444024698248, + "loss": 3.2110159397125244, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.146673202514648, + "eval_runtime": 58.7379, + "eval_samples_per_second": 41.574, + "eval_steps_per_second": 1.311, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.02919941022992134, + "learning_rate": 0.00024003388835314981, + "loss": 3.1533267498016357, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.03379782289266586, + "learning_rate": 0.00023977338718997595, + "loss": 3.1436538696289062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.033698830753564835, + "learning_rate": 0.00023951293696734195, + "loss": 3.15297532081604, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.03077736124396324, + "learning_rate": 0.00023925253789508806, + "loss": 3.128848075866699, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.031112324446439743, + "learning_rate": 0.00023899219018301286, + "loss": 3.1461119651794434, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.03346852958202362, + "learning_rate": 0.00023873189404087405, + "loss": 3.113943099975586, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.03307225927710533, + "learning_rate": 0.00023847164967838726, + "loss": 3.147498607635498, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.030230754986405373, + "learning_rate": 0.00023821145730522692, + "loss": 3.1162924766540527, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.02982981689274311, + "learning_rate": 0.0002379513171310251, + "loss": 3.144481658935547, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.031735632568597794, + "learning_rate": 0.00023769122936537207, + "loss": 3.133862018585205, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.0312766470015049, + "learning_rate": 0.00023743119421781614, + "loss": 3.110287666320801, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.029902562499046326, + "learning_rate": 0.00023717121189786256, + "loss": 3.1384904384613037, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.0306564774364233, + "learning_rate": 0.00023691128261497463, + "loss": 3.1138200759887695, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.03084711916744709, + "learning_rate": 0.0002366514065785725, + "loss": 3.155240535736084, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.03183220326900482, + "learning_rate": 0.00023639158399803382, + "loss": 3.1593515872955322, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.031826410442590714, + "learning_rate": 0.0002361318150826927, + "loss": 3.1517653465270996, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.030141707509756088, + "learning_rate": 0.0002358721000418404, + "loss": 3.1291589736938477, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.03158571943640709, + "learning_rate": 0.00023561243908472472, + "loss": 3.178101062774658, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.03510458394885063, + "learning_rate": 0.00023535283242054954, + "loss": 3.138767719268799, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.03660859912633896, + "learning_rate": 0.00023509328025847565, + "loss": 3.1708762645721436, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.034004807472229004, + "learning_rate": 0.00023483378280761907, + "loss": 3.1404073238372803, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.03077496401965618, + "learning_rate": 0.0002345743402770527, + "loss": 3.159081220626831, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.030384650453925133, + "learning_rate": 0.0002343149528758042, + "loss": 3.1872847080230713, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.02957277186214924, + "learning_rate": 0.00023405562081285766, + "loss": 3.165973424911499, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.029529055580496788, + "learning_rate": 0.0002337963442971522, + "loss": 3.123840808868408, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.027776138857007027, + "learning_rate": 0.00023353712353758212, + "loss": 3.1705269813537598, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.028943249955773354, + "learning_rate": 0.00023327795874299704, + "loss": 3.1181654930114746, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.028178401291370392, + "learning_rate": 0.0002330188501222012, + "loss": 3.201678991317749, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.02865665778517723, + "learning_rate": 0.0002327597978839538, + "loss": 3.18479061126709, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.029173649847507477, + "learning_rate": 0.00023250080223696852, + "loss": 3.1844916343688965, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.0280170775949955, + "learning_rate": 0.00023224186338991352, + "loss": 3.187105417251587, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.029103878885507584, + "learning_rate": 0.0002319829815514112, + "loss": 3.183281660079956, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.029045486822724342, + "learning_rate": 0.00023172415693003789, + "loss": 3.1848196983337402, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.028364380821585655, + "learning_rate": 0.000231465389734324, + "loss": 3.197348117828369, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.029651399701833725, + "learning_rate": 0.00023120668017275346, + "loss": 3.1429920196533203, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.0281438659876585, + "learning_rate": 0.00023094802845376403, + "loss": 3.1916513442993164, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.029437677934765816, + "learning_rate": 0.00023068943478574662, + "loss": 3.172391176223755, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.03433044254779816, + "learning_rate": 0.00023043089937704541, + "loss": 3.2296864986419678, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.03259492665529251, + "learning_rate": 0.00023017242243595796, + "loss": 3.1962363719940186, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.0299630519002676, + "learning_rate": 0.00022991400417073427, + "loss": 3.169673442840576, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.028495769947767258, + "learning_rate": 0.00022965564478957739, + "loss": 3.1855409145355225, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.0290420800447464, + "learning_rate": 0.0002293973445006427, + "loss": 3.23667573928833, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.03340664133429527, + "learning_rate": 0.00022913910351203816, + "loss": 3.1873788833618164, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.03253626823425293, + "learning_rate": 0.00022888092203182377, + "loss": 3.204796552658081, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.028142794966697693, + "learning_rate": 0.00022862280026801182, + "loss": 3.193922281265259, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.028805622830986977, + "learning_rate": 0.00022836473842856623, + "loss": 3.1863901615142822, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.030033612623810768, + "learning_rate": 0.00022810673672140272, + "loss": 3.198482036590576, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.028154220432043076, + "learning_rate": 0.0002278487953543889, + "loss": 3.215069055557251, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.028098132461309433, + "learning_rate": 0.00022759091453534308, + "loss": 3.2197628021240234, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.028679659590125084, + "learning_rate": 0.00022733309447203553, + "loss": 3.2062196731567383, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.029104258865118027, + "learning_rate": 0.00022707533537218695, + "loss": 3.2182908058166504, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.029629778116941452, + "learning_rate": 0.00022681763744346938, + "loss": 3.2165639400482178, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.028304366394877434, + "learning_rate": 0.00022656000089350524, + "loss": 3.1783785820007324, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.026900488883256912, + "learning_rate": 0.00022630242592986766, + "loss": 3.1876602172851562, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.027047287672758102, + "learning_rate": 0.0002260449127600803, + "loss": 3.1879465579986572, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.027424108237028122, + "learning_rate": 0.0002257874615916166, + "loss": 3.2407851219177246, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.029319508001208305, + "learning_rate": 0.00022553007263190067, + "loss": 3.1980130672454834, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.027114372700452805, + "learning_rate": 0.00022527274608830574, + "loss": 3.211357593536377, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.02579840086400509, + "learning_rate": 0.0002250154821681555, + "loss": 3.2067861557006836, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.02767331339418888, + "learning_rate": 0.0002247582810787227, + "loss": 3.198587417602539, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.026645265519618988, + "learning_rate": 0.00022450114302722957, + "loss": 3.177441120147705, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.026181906461715698, + "learning_rate": 0.0002242440682208477, + "loss": 3.226815700531006, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.027218692004680634, + "learning_rate": 0.00022398705686669748, + "loss": 3.1731510162353516, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.027551919221878052, + "learning_rate": 0.0002237301091718485, + "loss": 3.2268621921539307, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.025782009586691856, + "learning_rate": 0.00022347322534331865, + "loss": 3.2052621841430664, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.024231620132923126, + "learning_rate": 0.00022321640558807488, + "loss": 3.201770544052124, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.02474607340991497, + "learning_rate": 0.00022295965011303188, + "loss": 3.1935834884643555, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.026299327611923218, + "learning_rate": 0.00022270295912505317, + "loss": 3.223275899887085, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.02439914271235466, + "learning_rate": 0.00022244633283095, + "loss": 3.2513954639434814, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.02650653012096882, + "learning_rate": 0.0002221897714374814, + "loss": 3.1914851665496826, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.024847298860549927, + "learning_rate": 0.00022193327515135445, + "loss": 3.2173104286193848, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.028475770726799965, + "learning_rate": 0.00022167684417922337, + "loss": 3.2184863090515137, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.168707847595215, + "eval_runtime": 57.4035, + "eval_samples_per_second": 42.541, + "eval_steps_per_second": 1.341, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.02837727963924408, + "learning_rate": 0.00022142047872769014, + "loss": 3.119239330291748, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.03562312573194504, + "learning_rate": 0.00022116417900330365, + "loss": 3.092813014984131, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.0344194732606411, + "learning_rate": 0.00022090794521255997, + "loss": 3.0802855491638184, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.03188958019018173, + "learning_rate": 0.00022065177756190214, + "loss": 3.1177961826324463, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.032181013375520706, + "learning_rate": 0.00022039567625771968, + "loss": 3.1343281269073486, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.031085725873708725, + "learning_rate": 0.0002201396415063489, + "loss": 3.118828296661377, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.031995322555303574, + "learning_rate": 0.00021988367351407227, + "loss": 3.110109329223633, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.02949182130396366, + "learning_rate": 0.00021962777248711872, + "loss": 3.0972743034362793, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.029299668967723846, + "learning_rate": 0.00021937193863166292, + "loss": 3.1257238388061523, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.030558617785573006, + "learning_rate": 0.00021911617215382564, + "loss": 3.1216022968292236, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.03181575983762741, + "learning_rate": 0.00021886047325967348, + "loss": 3.093721866607666, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.030934303998947144, + "learning_rate": 0.00021860484215521805, + "loss": 3.0910396575927734, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.030235569924116135, + "learning_rate": 0.00021834927904641708, + "loss": 3.133371114730835, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.03350163996219635, + "learning_rate": 0.00021809378413917296, + "loss": 3.1178882122039795, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.03468719869852066, + "learning_rate": 0.00021783835763933344, + "loss": 3.1222469806671143, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.034203238785266876, + "learning_rate": 0.0002175829997526909, + "loss": 3.12453031539917, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.03321516141295433, + "learning_rate": 0.0002173277106849826, + "loss": 3.115428924560547, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.03298434242606163, + "learning_rate": 0.0002170724906418905, + "loss": 3.127288341522217, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.029987169429659843, + "learning_rate": 0.00021681733982904046, + "loss": 3.1194515228271484, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.031271349638700485, + "learning_rate": 0.00021656225845200332, + "loss": 3.1284165382385254, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.032201699912548065, + "learning_rate": 0.0002163072467162931, + "loss": 3.1912736892700195, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.0299740731716156, + "learning_rate": 0.00021605230482736845, + "loss": 3.1233785152435303, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.029613154008984566, + "learning_rate": 0.00021579743299063122, + "loss": 3.1535067558288574, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.03303765505552292, + "learning_rate": 0.00021554263141142726, + "loss": 3.1550910472869873, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.032644059509038925, + "learning_rate": 0.00021528790029504548, + "loss": 3.1367573738098145, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.031695179641246796, + "learning_rate": 0.00021503323984671814, + "loss": 3.1350858211517334, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.030685821548104286, + "learning_rate": 0.00021477865027162067, + "loss": 3.121593952178955, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.03091946430504322, + "learning_rate": 0.0002145241317748711, + "loss": 3.113060235977173, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.0282905176281929, + "learning_rate": 0.0002142696845615306, + "loss": 3.184551239013672, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.030020160600543022, + "learning_rate": 0.0002140153088366024, + "loss": 3.1673004627227783, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.031114062294363976, + "learning_rate": 0.00021376100480503276, + "loss": 3.162588596343994, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.03007666766643524, + "learning_rate": 0.00021350677267170973, + "loss": 3.132836103439331, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.029013793915510178, + "learning_rate": 0.00021325261264146342, + "loss": 3.1552720069885254, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.02784561738371849, + "learning_rate": 0.00021299852491906616, + "loss": 3.189272165298462, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.02759280428290367, + "learning_rate": 0.00021274450970923167, + "loss": 3.141814708709717, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.02699868381023407, + "learning_rate": 0.0002124905672166156, + "loss": 3.1127257347106934, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.02782190404832363, + "learning_rate": 0.0002122366976458146, + "loss": 3.193972587585449, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.027298318222165108, + "learning_rate": 0.00021198290120136702, + "loss": 3.1393446922302246, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.027321161702275276, + "learning_rate": 0.0002117291780877519, + "loss": 3.0926666259765625, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.026399722322821617, + "learning_rate": 0.00021147552850938941, + "loss": 3.1578259468078613, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.02770969271659851, + "learning_rate": 0.0002112219526706406, + "loss": 3.1816914081573486, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.028344258666038513, + "learning_rate": 0.00021096845077580675, + "loss": 3.1860158443450928, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.028091782703995705, + "learning_rate": 0.0002107150230291299, + "loss": 3.1585092544555664, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.026564039289951324, + "learning_rate": 0.00021046166963479204, + "loss": 3.131692409515381, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.026349127292633057, + "learning_rate": 0.00021020839079691553, + "loss": 3.160431385040283, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.030081138014793396, + "learning_rate": 0.00020995518671956256, + "loss": 3.1647543907165527, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.029250919818878174, + "learning_rate": 0.00020970205760673493, + "loss": 3.1522226333618164, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.0282477755099535, + "learning_rate": 0.00020944900366237453, + "loss": 3.139249324798584, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.0284534003585577, + "learning_rate": 0.00020919602509036195, + "loss": 3.154621124267578, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.02677394263446331, + "learning_rate": 0.0002089431220945178, + "loss": 3.1375081539154053, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.026511570438742638, + "learning_rate": 0.00020869029487860126, + "loss": 3.1355676651000977, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.02766270563006401, + "learning_rate": 0.00020843754364631068, + "loss": 3.143874168395996, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.027013977989554405, + "learning_rate": 0.0002081848686012832, + "loss": 3.1602487564086914, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.026629449799656868, + "learning_rate": 0.00020793226994709446, + "loss": 3.1549456119537354, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.0272684209048748, + "learning_rate": 0.0002076797478872588, + "loss": 3.1739020347595215, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.027645915746688843, + "learning_rate": 0.00020742730262522844, + "loss": 3.172959327697754, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.025413552299141884, + "learning_rate": 0.00020717493436439433, + "loss": 3.1649041175842285, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.028406929224729538, + "learning_rate": 0.00020692264330808455, + "loss": 3.1821887493133545, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.027753977105021477, + "learning_rate": 0.00020667042965956587, + "loss": 3.163238048553467, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.02693209983408451, + "learning_rate": 0.00020641829362204195, + "loss": 3.1921019554138184, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.02671878971159458, + "learning_rate": 0.00020616623539865446, + "loss": 3.159438133239746, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.027099428698420525, + "learning_rate": 0.00020591425519248213, + "loss": 3.1806793212890625, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.025098824873566628, + "learning_rate": 0.00020566235320654069, + "loss": 3.158310651779175, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.02655479498207569, + "learning_rate": 0.00020541052964378323, + "loss": 3.1927309036254883, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.025836164131760597, + "learning_rate": 0.00020515878470709928, + "loss": 3.1578540802001953, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.02657160349190235, + "learning_rate": 0.00020490711859931545, + "loss": 3.1658592224121094, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.027223270386457443, + "learning_rate": 0.00020465553152319418, + "loss": 3.1760241985321045, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.025695348158478737, + "learning_rate": 0.00020440402368143492, + "loss": 3.202389717102051, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.025819813832640648, + "learning_rate": 0.00020415259527667299, + "loss": 3.2111103534698486, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.02621106244623661, + "learning_rate": 0.0002039012465114796, + "loss": 3.1560299396514893, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.026585502550005913, + "learning_rate": 0.00020364997758836193, + "loss": 3.176602840423584, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.03038652054965496, + "learning_rate": 0.00020339878870976287, + "loss": 3.1799943447113037, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.193845748901367, + "eval_runtime": 57.9911, + "eval_samples_per_second": 42.11, + "eval_steps_per_second": 1.328, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.03135736659169197, + "learning_rate": 0.00020314768007806077, + "loss": 3.070986270904541, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.038986921310424805, + "learning_rate": 0.00020289665189556923, + "loss": 3.0708835124969482, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.035115309059619904, + "learning_rate": 0.00020264570436453707, + "loss": 3.1070001125335693, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.032868221402168274, + "learning_rate": 0.00020239483768714843, + "loss": 3.082258701324463, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.03703184798359871, + "learning_rate": 0.0002021440520655219, + "loss": 3.057744264602661, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.03831101953983307, + "learning_rate": 0.00020189334770171098, + "loss": 3.0789475440979004, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.036301176995038986, + "learning_rate": 0.00020164272479770352, + "loss": 3.0565457344055176, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.03166409209370613, + "learning_rate": 0.00020139218355542203, + "loss": 3.0574910640716553, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.02984458953142166, + "learning_rate": 0.000201141724176723, + "loss": 3.080018997192383, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.034001126885414124, + "learning_rate": 0.0002008913468633969, + "loss": 3.0813302993774414, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.03814122825860977, + "learning_rate": 0.00020064105181716847, + "loss": 3.0579185485839844, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.03870106488466263, + "learning_rate": 0.00020039083923969564, + "loss": 3.0857529640197754, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.04005642607808113, + "learning_rate": 0.0002001407093325704, + "loss": 3.0880727767944336, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.03760755434632301, + "learning_rate": 0.0001998906622973177, + "loss": 3.0795767307281494, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.03260907158255577, + "learning_rate": 0.0001996406983353961, + "loss": 3.1006722450256348, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.029866183176636696, + "learning_rate": 0.0001993908176481969, + "loss": 3.0850343704223633, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.030988436192274094, + "learning_rate": 0.0001991410204370445, + "loss": 3.0743892192840576, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.03698134422302246, + "learning_rate": 0.00019889130690319608, + "loss": 3.056002378463745, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.03696191683411598, + "learning_rate": 0.00019864167724784117, + "loss": 3.069124698638916, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.03453022986650467, + "learning_rate": 0.00019839213167210213, + "loss": 3.1017374992370605, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.032458193600177765, + "learning_rate": 0.00019814267037703294, + "loss": 3.1166419982910156, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.02892368845641613, + "learning_rate": 0.0001978932935636205, + "loss": 3.0863287448883057, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.030625877901911736, + "learning_rate": 0.00019764400143278276, + "loss": 3.0938801765441895, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.030280394479632378, + "learning_rate": 0.00019739479418537018, + "loss": 3.086432933807373, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.029181072488427162, + "learning_rate": 0.00019714567202216453, + "loss": 3.0977859497070312, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.028039127588272095, + "learning_rate": 0.00019689663514387896, + "loss": 3.112367868423462, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.02910272404551506, + "learning_rate": 0.00019664768375115798, + "loss": 3.1253533363342285, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.03086872771382332, + "learning_rate": 0.00019639881804457723, + "loss": 3.103689670562744, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.02883610874414444, + "learning_rate": 0.00019615003822464342, + "loss": 3.111599922180176, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.028490116819739342, + "learning_rate": 0.00019590134449179371, + "loss": 3.1410303115844727, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.028714407235383987, + "learning_rate": 0.00019565273704639648, + "loss": 3.133023738861084, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.029553238302469254, + "learning_rate": 0.00019540421608874994, + "loss": 3.107003688812256, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.02997349575161934, + "learning_rate": 0.0001951557818190831, + "loss": 3.135680675506592, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.029087591916322708, + "learning_rate": 0.00019490743443755504, + "loss": 3.1103882789611816, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.029947761446237564, + "learning_rate": 0.0001946591741442546, + "loss": 3.0961623191833496, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.02950555644929409, + "learning_rate": 0.00019441100113920076, + "loss": 3.1162495613098145, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.029569562524557114, + "learning_rate": 0.00019416291562234192, + "loss": 3.096372127532959, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.031366728246212006, + "learning_rate": 0.00019391491779355608, + "loss": 3.120089530944824, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.0313473716378212, + "learning_rate": 0.00019366700785265065, + "loss": 3.100785732269287, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.028021851554512978, + "learning_rate": 0.0001934191859993622, + "loss": 3.0940558910369873, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.027477117255330086, + "learning_rate": 0.00019317145243335642, + "loss": 3.1210484504699707, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.029690749943256378, + "learning_rate": 0.0001929238073542276, + "loss": 3.1350059509277344, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.029659733176231384, + "learning_rate": 0.00019267625096149903, + "loss": 3.1220788955688477, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.028260033577680588, + "learning_rate": 0.00019242878345462227, + "loss": 3.1000189781188965, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.027554461732506752, + "learning_rate": 0.00019218140503297765, + "loss": 3.1044349670410156, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.027929967269301414, + "learning_rate": 0.00019193411589587332, + "loss": 3.1353745460510254, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.029309319332242012, + "learning_rate": 0.00019168691624254564, + "loss": 3.1578292846679688, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.029183287173509598, + "learning_rate": 0.00019143980627215915, + "loss": 3.1217150688171387, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.02707512304186821, + "learning_rate": 0.00019119278618380555, + "loss": 3.119617223739624, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.03006536327302456, + "learning_rate": 0.00019094585617650476, + "loss": 3.1481480598449707, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.03026784397661686, + "learning_rate": 0.00019069901644920366, + "loss": 3.142770767211914, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.02827967144548893, + "learning_rate": 0.00019045226720077667, + "loss": 3.1582541465759277, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.028892774134874344, + "learning_rate": 0.0001902056086300251, + "loss": 3.1530709266662598, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.03023783676326275, + "learning_rate": 0.0001899590409356773, + "loss": 3.1102521419525146, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.027951840311288834, + "learning_rate": 0.00018971256431638854, + "loss": 3.139113426208496, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.02807226963341236, + "learning_rate": 0.00018946617897074041, + "loss": 3.149428367614746, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.029721451923251152, + "learning_rate": 0.0001892198850972413, + "loss": 3.120518207550049, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.02867680788040161, + "learning_rate": 0.00018897368289432553, + "loss": 3.1109931468963623, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.02663845382630825, + "learning_rate": 0.00018872757256035403, + "loss": 3.116053581237793, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.028366081416606903, + "learning_rate": 0.00018848155429361323, + "loss": 3.1444268226623535, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.028931576758623123, + "learning_rate": 0.00018823562829231567, + "loss": 3.163870096206665, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.02664567157626152, + "learning_rate": 0.0001879897947545996, + "loss": 3.1349146366119385, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.028031354770064354, + "learning_rate": 0.00018774405387852848, + "loss": 3.1638219356536865, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.029421644285321236, + "learning_rate": 0.00018749840586209144, + "loss": 3.1601340770721436, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.027983693405985832, + "learning_rate": 0.00018725285090320245, + "loss": 3.1636788845062256, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.02719871513545513, + "learning_rate": 0.00018700738919970103, + "loss": 3.1246869564056396, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.03030397929251194, + "learning_rate": 0.00018676202094935085, + "loss": 3.1749579906463623, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.029340893030166626, + "learning_rate": 0.000186516746349841, + "loss": 3.1751790046691895, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.027624106034636497, + "learning_rate": 0.0001862715655987847, + "loss": 3.178704023361206, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.02830546908080578, + "learning_rate": 0.00018602647889371957, + "loss": 3.177117347717285, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.03168068826198578, + "learning_rate": 0.00018578148643210764, + "loss": 3.125427722930908, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.033327918499708176, + "learning_rate": 0.00018553658841133483, + "loss": 3.13071870803833, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.218388080596924, + "eval_runtime": 57.611, + "eval_samples_per_second": 42.388, + "eval_steps_per_second": 1.337, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.03336749225854874, + "learning_rate": 0.00018529178502871117, + "loss": 3.064338445663452, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.038479577749967575, + "learning_rate": 0.0001850470764814702, + "loss": 3.0478687286376953, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.032562077045440674, + "learning_rate": 0.0001848024629667691, + "loss": 3.019803762435913, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.0331728458404541, + "learning_rate": 0.00018455794468168887, + "loss": 3.0599141120910645, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.03361169621348381, + "learning_rate": 0.00018431352182323322, + "loss": 3.0315823554992676, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.032973092049360275, + "learning_rate": 0.0001840691945883293, + "loss": 3.025928258895874, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.030388766899704933, + "learning_rate": 0.00018382496317382715, + "loss": 3.070671558380127, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.030899515375494957, + "learning_rate": 0.0001835808277764997, + "loss": 3.012938976287842, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.0335412323474884, + "learning_rate": 0.00018333678859304222, + "loss": 3.0156331062316895, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.035731345415115356, + "learning_rate": 0.00018309284582007274, + "loss": 3.0578935146331787, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.03379053995013237, + "learning_rate": 0.0001828489996541318, + "loss": 3.048701286315918, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.031242679804563522, + "learning_rate": 0.00018260525029168147, + "loss": 3.0832977294921875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.030997611582279205, + "learning_rate": 0.00018236159792910648, + "loss": 3.0663583278656006, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.03397331386804581, + "learning_rate": 0.00018211804276271304, + "loss": 3.0115303993225098, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.03349050134420395, + "learning_rate": 0.00018187458498872914, + "loss": 3.02864933013916, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.031079065054655075, + "learning_rate": 0.00018163122480330433, + "loss": 3.053401470184326, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.029346197843551636, + "learning_rate": 0.00018138796240250955, + "loss": 3.0700926780700684, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.030784208327531815, + "learning_rate": 0.00018114479798233686, + "loss": 3.066525936126709, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.03368140384554863, + "learning_rate": 0.00018090173173869939, + "loss": 3.0956242084503174, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.03356792405247688, + "learning_rate": 0.00018065876386743143, + "loss": 3.0394320487976074, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.029909325763583183, + "learning_rate": 0.00018041589456428754, + "loss": 3.031525135040283, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.030711578205227852, + "learning_rate": 0.0001801731240249434, + "loss": 3.075474977493286, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.03452072665095329, + "learning_rate": 0.00017993045244499463, + "loss": 3.0228147506713867, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.03235075995326042, + "learning_rate": 0.00017968788001995742, + "loss": 3.0664501190185547, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.029671598225831985, + "learning_rate": 0.000179445406945268, + "loss": 3.055302143096924, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.033868469297885895, + "learning_rate": 0.0001792030334162825, + "loss": 3.053016185760498, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.0329158715903759, + "learning_rate": 0.00017896075962827706, + "loss": 3.1017332077026367, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.031983695924282074, + "learning_rate": 0.00017871858577644712, + "loss": 3.0706472396850586, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.02963697351515293, + "learning_rate": 0.0001784765120559078, + "loss": 3.0711426734924316, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.03162391856312752, + "learning_rate": 0.00017823453866169355, + "loss": 3.044867515563965, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.030773654580116272, + "learning_rate": 0.00017799266578875808, + "loss": 3.0963196754455566, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.031220417469739914, + "learning_rate": 0.00017775089363197371, + "loss": 3.0892245769500732, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.030928561463952065, + "learning_rate": 0.00017750922238613198, + "loss": 3.1181323528289795, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.03192073106765747, + "learning_rate": 0.00017726765224594342, + "loss": 3.066584348678589, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.03075495734810829, + "learning_rate": 0.0001770261834060363, + "loss": 3.0436511039733887, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.02895325981080532, + "learning_rate": 0.00017678481606095786, + "loss": 3.123373031616211, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.029738714918494225, + "learning_rate": 0.0001765435504051734, + "loss": 3.1049060821533203, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.031937118619680405, + "learning_rate": 0.0001763023866330664, + "loss": 3.099111318588257, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.029886579141020775, + "learning_rate": 0.00017606132493893793, + "loss": 3.0970301628112793, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.029569782316684723, + "learning_rate": 0.00017582036551700724, + "loss": 3.062499761581421, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.02924283780157566, + "learning_rate": 0.00017557950856141085, + "loss": 3.0618538856506348, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.030955800786614418, + "learning_rate": 0.00017533875426620286, + "loss": 3.0745606422424316, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.03145027533173561, + "learning_rate": 0.00017509810282535497, + "loss": 3.122938632965088, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.028926467522978783, + "learning_rate": 0.00017485755443275513, + "loss": 3.0808773040771484, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.029592076316475868, + "learning_rate": 0.0001746171092822094, + "loss": 3.1167049407958984, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.028070243075489998, + "learning_rate": 0.00017437676756743974, + "loss": 3.0781636238098145, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.028642063960433006, + "learning_rate": 0.00017413652948208524, + "loss": 3.093937397003174, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.0290757454931736, + "learning_rate": 0.00017389639521970136, + "loss": 3.1079559326171875, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.027867283672094345, + "learning_rate": 0.00017365636497376004, + "loss": 3.0885753631591797, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.02896495908498764, + "learning_rate": 0.00017341643893764927, + "loss": 3.0889246463775635, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.02777806483209133, + "learning_rate": 0.000173176617304673, + "loss": 3.084439277648926, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.02904277853667736, + "learning_rate": 0.00017293690026805124, + "loss": 3.125246047973633, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.029010627418756485, + "learning_rate": 0.00017269728802091975, + "loss": 3.111422061920166, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.028178894892334938, + "learning_rate": 0.00017245778075632975, + "loss": 3.115067481994629, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.02763799950480461, + "learning_rate": 0.00017221837866724792, + "loss": 3.1228506565093994, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.028052521869540215, + "learning_rate": 0.00017197908194655615, + "loss": 3.1150546073913574, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.029438253492116928, + "learning_rate": 0.00017173989078705175, + "loss": 3.0921788215637207, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.028195543214678764, + "learning_rate": 0.0001715008053814463, + "loss": 3.1408450603485107, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.02937602624297142, + "learning_rate": 0.00017126182592236682, + "loss": 3.111612319946289, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.028430236503481865, + "learning_rate": 0.00017102295260235472, + "loss": 3.093441963195801, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.027624569833278656, + "learning_rate": 0.0001707841856138659, + "loss": 3.1184372901916504, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.029797039926052094, + "learning_rate": 0.00017054552514927061, + "loss": 3.1074559688568115, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.028728218749165535, + "learning_rate": 0.00017030697140085334, + "loss": 3.161461114883423, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.02704412117600441, + "learning_rate": 0.0001700685245608126, + "loss": 3.13325834274292, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.028067929670214653, + "learning_rate": 0.00016983018482126047, + "loss": 3.093172550201416, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.028848597779870033, + "learning_rate": 0.00016959195237422305, + "loss": 3.1461825370788574, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.02956785447895527, + "learning_rate": 0.00016935382741164, + "loss": 3.141045570373535, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.028214773163199425, + "learning_rate": 0.00016911581012536424, + "loss": 3.136033773422241, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.029155246913433075, + "learning_rate": 0.0001688779007071622, + "loss": 3.0986382961273193, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.02968468703329563, + "learning_rate": 0.00016864009934871275, + "loss": 3.11630916595459, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.028356682509183884, + "learning_rate": 0.0001684024062416087, + "loss": 3.1134822368621826, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.03301718086004257, + "learning_rate": 0.0001681648215773546, + "loss": 3.118591785430908, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.246211528778076, + "eval_runtime": 57.5155, + "eval_samples_per_second": 42.458, + "eval_steps_per_second": 1.339, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.033637888729572296, + "learning_rate": 0.0001679273455473684, + "loss": 3.0192506313323975, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.04057888314127922, + "learning_rate": 0.0001676899783429802, + "loss": 2.977440595626831, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.03579854592680931, + "learning_rate": 0.00016745272015543242, + "loss": 2.9902219772338867, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.03131850063800812, + "learning_rate": 0.0001672155711758799, + "loss": 2.9988656044006348, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.03262636810541153, + "learning_rate": 0.00016697853159538896, + "loss": 3.037747621536255, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.03260403499007225, + "learning_rate": 0.0001667416016049383, + "loss": 2.987508773803711, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.033165134489536285, + "learning_rate": 0.0001665047813954181, + "loss": 3.0511183738708496, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.03303489834070206, + "learning_rate": 0.00016626807115763023, + "loss": 3.0083041191101074, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.03185006603598595, + "learning_rate": 0.00016603147108228754, + "loss": 2.9966228008270264, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.03135651350021362, + "learning_rate": 0.00016579498136001485, + "loss": 3.036759376525879, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.03216925635933876, + "learning_rate": 0.00016555860218134738, + "loss": 3.0304150581359863, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.03231671452522278, + "learning_rate": 0.00016532233373673164, + "loss": 3.0300474166870117, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.03263348713517189, + "learning_rate": 0.00016508617621652488, + "loss": 3.0298163890838623, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.032594650983810425, + "learning_rate": 0.00016485012981099488, + "loss": 2.98789119720459, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.03406650573015213, + "learning_rate": 0.0001646141947103201, + "loss": 3.0102760791778564, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.03379780799150467, + "learning_rate": 0.0001643783711045889, + "loss": 3.0564732551574707, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.03172587603330612, + "learning_rate": 0.0001641426591838005, + "loss": 3.0410943031311035, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.02938298135995865, + "learning_rate": 0.00016390705913786344, + "loss": 3.0517544746398926, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.03206252679228783, + "learning_rate": 0.00016367157115659646, + "loss": 3.0352776050567627, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.033386752009391785, + "learning_rate": 0.00016343619542972808, + "loss": 3.020535945892334, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.033984702080488205, + "learning_rate": 0.0001632009321468962, + "loss": 3.0552730560302734, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.0329931266605854, + "learning_rate": 0.00016296578149764833, + "loss": 3.0075573921203613, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.032101184129714966, + "learning_rate": 0.00016273074367144087, + "loss": 3.046388864517212, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.032873060554265976, + "learning_rate": 0.00016249581885763968, + "loss": 3.0340287685394287, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.030672110617160797, + "learning_rate": 0.0001622610072455194, + "loss": 3.0347838401794434, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.030014054849743843, + "learning_rate": 0.00016202630902426355, + "loss": 3.0315332412719727, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.02932834066450596, + "learning_rate": 0.00016179172438296423, + "loss": 3.084340810775757, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.029181934893131256, + "learning_rate": 0.00016155725351062205, + "loss": 3.032886028289795, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.02966221794486046, + "learning_rate": 0.00016132289659614605, + "loss": 3.032766819000244, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.029713116586208344, + "learning_rate": 0.00016108865382835318, + "loss": 3.028778076171875, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.029243769124150276, + "learning_rate": 0.00016085452539596864, + "loss": 3.0492146015167236, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.02997511625289917, + "learning_rate": 0.0001606205114876256, + "loss": 3.0772786140441895, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.02930292673408985, + "learning_rate": 0.00016038661229186472, + "loss": 3.055750608444214, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.03037521243095398, + "learning_rate": 0.00016015282799713453, + "loss": 3.033710479736328, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.029957240447402, + "learning_rate": 0.0001599191587917905, + "loss": 3.0394721031188965, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.02925589308142662, + "learning_rate": 0.0001596856048640961, + "loss": 3.0878353118896484, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.03003518097102642, + "learning_rate": 0.00015945216640222116, + "loss": 3.048628330230713, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.031545430421829224, + "learning_rate": 0.000159218843594243, + "loss": 3.0775623321533203, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.031310439109802246, + "learning_rate": 0.0001589856366281455, + "loss": 3.01645565032959, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.03074057213962078, + "learning_rate": 0.0001587525456918194, + "loss": 3.0651440620422363, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.028704499825835228, + "learning_rate": 0.0001585195709730619, + "loss": 3.0527148246765137, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.029061976820230484, + "learning_rate": 0.00015828671265957623, + "loss": 3.0854501724243164, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.03034680150449276, + "learning_rate": 0.00015805397093897255, + "loss": 3.056356906890869, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.028715113177895546, + "learning_rate": 0.00015782134599876633, + "loss": 3.0874247550964355, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.030256496742367744, + "learning_rate": 0.00015758883802637953, + "loss": 3.052424430847168, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.02949833683669567, + "learning_rate": 0.00015735644720913923, + "loss": 3.0663650035858154, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.028980428352952003, + "learning_rate": 0.00015712417373427897, + "loss": 3.0700113773345947, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.028526026755571365, + "learning_rate": 0.00015689201778893713, + "loss": 3.0856711864471436, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.029572727158665657, + "learning_rate": 0.00015665997956015743, + "loss": 3.0862035751342773, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.02955584041774273, + "learning_rate": 0.00015642805923488895, + "loss": 3.030802011489868, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.02881692908704281, + "learning_rate": 0.00015619625699998585, + "loss": 3.074392318725586, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.029564999043941498, + "learning_rate": 0.00015596457304220692, + "loss": 3.057894229888916, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.030315004289150238, + "learning_rate": 0.00015573300754821547, + "loss": 3.096233367919922, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.030666006729006767, + "learning_rate": 0.00015550156070458002, + "loss": 3.040799617767334, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.028901133686304092, + "learning_rate": 0.0001552702326977731, + "loss": 3.055051803588867, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.030359387397766113, + "learning_rate": 0.00015503902371417122, + "loss": 3.085972309112549, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.03117106668651104, + "learning_rate": 0.00015480793394005547, + "loss": 3.0786352157592773, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.029238386079669, + "learning_rate": 0.0001545769635616107, + "loss": 3.0375771522521973, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.029324093833565712, + "learning_rate": 0.0001543461127649256, + "loss": 3.1140198707580566, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.02875387854874134, + "learning_rate": 0.00015411538173599231, + "loss": 3.090881109237671, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.029023943468928337, + "learning_rate": 0.00015388477066070674, + "loss": 3.0755701065063477, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.0291463453322649, + "learning_rate": 0.00015365427972486827, + "loss": 3.0723776817321777, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.02795177884399891, + "learning_rate": 0.00015342390911417904, + "loss": 3.058993339538574, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.029499078169465065, + "learning_rate": 0.00015319365901424452, + "loss": 3.0856404304504395, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.029889609664678574, + "learning_rate": 0.0001529635296105732, + "loss": 3.0622942447662354, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.03107610158622265, + "learning_rate": 0.0001527335210885761, + "loss": 3.114957332611084, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.02967257983982563, + "learning_rate": 0.00015250363363356685, + "loss": 3.088871479034424, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.029068559408187866, + "learning_rate": 0.00015227386743076155, + "loss": 3.1103081703186035, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.029334546998143196, + "learning_rate": 0.00015204422266527911, + "loss": 3.0841622352600098, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.029259610921144485, + "learning_rate": 0.00015181469952213968, + "loss": 3.0662848949432373, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.030072513967752457, + "learning_rate": 0.0001515852981862663, + "loss": 3.0900328159332275, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.033321622759103775, + "learning_rate": 0.00015135601884248303, + "loss": 3.0871129035949707, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.271463394165039, + "eval_runtime": 57.9048, + "eval_samples_per_second": 42.173, + "eval_steps_per_second": 1.33, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.03635408729314804, + "learning_rate": 0.0001511268616755166, + "loss": 2.9576284885406494, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.03927835449576378, + "learning_rate": 0.00015089782686999445, + "loss": 2.999624729156494, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.03371119126677513, + "learning_rate": 0.00015066891461044588, + "loss": 2.971052646636963, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.033049341291189194, + "learning_rate": 0.00015044012508130138, + "loss": 2.980811595916748, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.03830729424953461, + "learning_rate": 0.00015021145846689254, + "loss": 2.9985947608947754, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.036573827266693115, + "learning_rate": 0.00014998291495145206, + "loss": 2.9892568588256836, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.031196987256407738, + "learning_rate": 0.00014975449471911294, + "loss": 2.9854254722595215, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.03376495838165283, + "learning_rate": 0.0001495261979539097, + "loss": 2.9840264320373535, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.036272983998060226, + "learning_rate": 0.0001492980248397766, + "loss": 3.0168871879577637, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.03328242897987366, + "learning_rate": 0.00014906997556054865, + "loss": 2.9815053939819336, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.03158426657319069, + "learning_rate": 0.00014884205029996113, + "loss": 2.994525909423828, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.03320910036563873, + "learning_rate": 0.00014861424924164925, + "loss": 3.008638858795166, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.03160254284739494, + "learning_rate": 0.00014838657256914832, + "loss": 2.979243040084839, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.03158991411328316, + "learning_rate": 0.0001481590204658931, + "loss": 2.938634157180786, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.031116485595703125, + "learning_rate": 0.00014793159311521833, + "loss": 2.9957656860351562, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.030757686123251915, + "learning_rate": 0.00014770429070035812, + "loss": 3.0106635093688965, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.02992827817797661, + "learning_rate": 0.00014747711340444601, + "loss": 2.996793746948242, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.0330309197306633, + "learning_rate": 0.00014725006141051434, + "loss": 2.9859161376953125, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.029488038271665573, + "learning_rate": 0.00014702313490149512, + "loss": 3.0081536769866943, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.03095114603638649, + "learning_rate": 0.00014679633406021892, + "loss": 3.006385087966919, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.03194110840559006, + "learning_rate": 0.00014656965906941494, + "loss": 3.0125250816345215, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.03306785598397255, + "learning_rate": 0.00014634311011171112, + "loss": 2.9640283584594727, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.030883541330695152, + "learning_rate": 0.00014611668736963396, + "loss": 3.0223653316497803, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.03224025294184685, + "learning_rate": 0.00014589039102560822, + "loss": 3.0274910926818848, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.03182433918118477, + "learning_rate": 0.00014566422126195656, + "loss": 3.047736406326294, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.03387221321463585, + "learning_rate": 0.00014543817826089992, + "loss": 2.9821343421936035, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.03407924994826317, + "learning_rate": 0.0001452122622045574, + "loss": 3.041207790374756, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.031546324491500854, + "learning_rate": 0.00014498647327494506, + "loss": 3.0298609733581543, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.03354770317673683, + "learning_rate": 0.00014476081165397715, + "loss": 3.0212411880493164, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.031692177057266235, + "learning_rate": 0.00014453527752346517, + "loss": 3.007350444793701, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.03218817710876465, + "learning_rate": 0.00014430987106511795, + "loss": 3.0019829273223877, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.03233771771192551, + "learning_rate": 0.00014408459246054126, + "loss": 3.0400614738464355, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.028998777270317078, + "learning_rate": 0.00014385944189123794, + "loss": 3.022221088409424, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.030422436073422432, + "learning_rate": 0.00014363441953860805, + "loss": 3.0071797370910645, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.030207227915525436, + "learning_rate": 0.00014340952558394778, + "loss": 2.9762964248657227, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.0305719543248415, + "learning_rate": 0.00014318476020845018, + "loss": 3.075176239013672, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.03173545375466347, + "learning_rate": 0.0001429601235932047, + "loss": 3.028982162475586, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.02799576334655285, + "learning_rate": 0.00014273561591919705, + "loss": 3.0242385864257812, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.03140150383114815, + "learning_rate": 0.00014251123736730884, + "loss": 3.012221336364746, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.029416466131806374, + "learning_rate": 0.00014228698811831787, + "loss": 3.0037949085235596, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.028385788202285767, + "learning_rate": 0.0001420628683528978, + "loss": 3.0350992679595947, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.031597770750522614, + "learning_rate": 0.00014183887825161779, + "loss": 3.065139055252075, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.02926146425306797, + "learning_rate": 0.0001416150179949428, + "loss": 3.014230728149414, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.02970231883227825, + "learning_rate": 0.0001413912877632326, + "loss": 3.06636381149292, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.029684338718652725, + "learning_rate": 0.00014116768773674306, + "loss": 3.024610996246338, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.02908361330628395, + "learning_rate": 0.00014094421809562438, + "loss": 3.0093719959259033, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.029891187325119972, + "learning_rate": 0.00014072087901992209, + "loss": 3.0735411643981934, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.028607673943042755, + "learning_rate": 0.00014049767068957645, + "loss": 3.044274091720581, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.029806701466441154, + "learning_rate": 0.00014027459328442242, + "loss": 3.0449328422546387, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.03161340951919556, + "learning_rate": 0.00014005164698418948, + "loss": 3.046813488006592, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.028860243037343025, + "learning_rate": 0.0001398288319685013, + "loss": 3.011828660964966, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.0285879448056221, + "learning_rate": 0.00013960614841687595, + "loss": 3.043640613555908, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.0296401996165514, + "learning_rate": 0.00013938359650872558, + "loss": 3.071589469909668, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.02807026542723179, + "learning_rate": 0.00013916117642335623, + "loss": 3.018320083618164, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.029037052765488625, + "learning_rate": 0.00013893888833996774, + "loss": 3.0471763610839844, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.029226379469037056, + "learning_rate": 0.00013871673243765362, + "loss": 3.037463426589966, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.029322804883122444, + "learning_rate": 0.00013849470889540098, + "loss": 3.0472755432128906, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.028253905475139618, + "learning_rate": 0.00013827281789208996, + "loss": 3.0265414714813232, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.028875533491373062, + "learning_rate": 0.00013805105960649423, + "loss": 3.063941717147827, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.028460513800382614, + "learning_rate": 0.00013782943421728042, + "loss": 3.0512683391571045, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.02951696142554283, + "learning_rate": 0.00013760794190300827, + "loss": 3.0528573989868164, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.030593490228056908, + "learning_rate": 0.00013738658284212973, + "loss": 3.057919979095459, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.028074532747268677, + "learning_rate": 0.00013716535721299016, + "loss": 3.0263280868530273, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.02955678664147854, + "learning_rate": 0.0001369442651938271, + "loss": 3.081322193145752, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.029724644497036934, + "learning_rate": 0.00013672330696277014, + "loss": 3.059583902359009, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.029713226482272148, + "learning_rate": 0.00013650248269784143, + "loss": 3.018678903579712, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.029758386313915253, + "learning_rate": 0.00013628179257695508, + "loss": 3.0379555225372314, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.028971998021006584, + "learning_rate": 0.00013606123677791727, + "loss": 3.041267156600952, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.02968725562095642, + "learning_rate": 0.00013584081547842565, + "loss": 3.0675506591796875, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.029876509681344032, + "learning_rate": 0.00013562052885606957, + "loss": 3.051586151123047, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.02853119559586048, + "learning_rate": 0.00013540037708833043, + "loss": 3.0335168838500977, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.03374241292476654, + "learning_rate": 0.00013518036035258012, + "loss": 3.027909278869629, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.030695006251335144, + "learning_rate": 0.0001349604788260823, + "loss": 2.9685583114624023, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0339728482067585, + "learning_rate": 0.00013474073268599152, + "loss": 2.936551570892334, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.03372679278254509, + "learning_rate": 0.00013452112210935339, + "loss": 2.9631872177124023, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.03172554448246956, + "learning_rate": 0.00013430164727310399, + "loss": 2.969510078430176, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.030516354367136955, + "learning_rate": 0.00013408230835407027, + "loss": 2.930342674255371, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.03261839970946312, + "learning_rate": 0.00013386310552896977, + "loss": 2.9920477867126465, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.030618635937571526, + "learning_rate": 0.00013364403897441016, + "loss": 2.9693596363067627, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.030576011165976524, + "learning_rate": 0.00013342510886688944, + "loss": 2.929804801940918, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.03224785998463631, + "learning_rate": 0.00013320631538279548, + "loss": 2.95280122756958, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.030605586245656013, + "learning_rate": 0.00013298765869840658, + "loss": 2.931358575820923, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.031206639483571053, + "learning_rate": 0.00013276913898989013, + "loss": 2.962998151779175, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.031317099928855896, + "learning_rate": 0.00013255075643330369, + "loss": 2.9706358909606934, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.031614501029253006, + "learning_rate": 0.00013233251120459407, + "loss": 2.9326977729797363, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.031002728268504143, + "learning_rate": 0.0001321144034795976, + "loss": 2.99625301361084, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.03216253221035004, + "learning_rate": 0.00013189643343403977, + "loss": 2.920912027359009, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.033380527049303055, + "learning_rate": 0.00013167860124353476, + "loss": 2.9688234329223633, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.03085758350789547, + "learning_rate": 0.00013146090708358657, + "loss": 2.9811222553253174, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.03258587792515755, + "learning_rate": 0.00013124335112958704, + "loss": 2.932366371154785, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.031639691442251205, + "learning_rate": 0.0001310259335568172, + "loss": 2.988865613937378, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.03144839406013489, + "learning_rate": 0.00013080865454044647, + "loss": 2.9750723838806152, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.030864672735333443, + "learning_rate": 0.00013059151425553263, + "loss": 2.999329090118408, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.030679022893309593, + "learning_rate": 0.00013037451287702183, + "loss": 2.974264144897461, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.03236627206206322, + "learning_rate": 0.0001301576505797479, + "loss": 2.9808404445648193, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.03115275874733925, + "learning_rate": 0.00012994092753843293, + "loss": 3.0216290950775146, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.029868317767977715, + "learning_rate": 0.00012972434392768687, + "loss": 2.9739654064178467, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.03446527197957039, + "learning_rate": 0.00012950789992200714, + "loss": 2.9565515518188477, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.033536117523908615, + "learning_rate": 0.00012929159569577886, + "loss": 2.958047389984131, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.03113192319869995, + "learning_rate": 0.0001290754314232743, + "loss": 2.9662246704101562, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.03089931420981884, + "learning_rate": 0.00012885940727865334, + "loss": 2.9972119331359863, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.030822310596704483, + "learning_rate": 0.00012864352343596247, + "loss": 3.020500421524048, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.032264843583106995, + "learning_rate": 0.0001284277800691355, + "loss": 2.9895882606506348, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.031226349994540215, + "learning_rate": 0.00012821217735199298, + "loss": 2.958217144012451, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.03110436722636223, + "learning_rate": 0.00012799671545824212, + "loss": 2.9884791374206543, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.03358910232782364, + "learning_rate": 0.0001277813945614768, + "loss": 2.9908103942871094, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.03348501771688461, + "learning_rate": 0.00012756621483517682, + "loss": 2.985677480697632, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.029595404863357544, + "learning_rate": 0.00012735117645270905, + "loss": 2.979318380355835, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.031459368765354156, + "learning_rate": 0.00012713627958732567, + "loss": 2.9921798706054688, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.03067564032971859, + "learning_rate": 0.00012692152441216539, + "loss": 2.977778434753418, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.030506785959005356, + "learning_rate": 0.0001267069111002525, + "loss": 2.9841768741607666, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.029978664591908455, + "learning_rate": 0.00012649243982449718, + "loss": 2.995443344116211, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.030900927260518074, + "learning_rate": 0.00012627811075769502, + "loss": 3.0177016258239746, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.029710259288549423, + "learning_rate": 0.00012606392407252687, + "loss": 3.01532244682312, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.031190430745482445, + "learning_rate": 0.00012584987994155943, + "loss": 2.9791834354400635, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.03150483965873718, + "learning_rate": 0.00012563597853724388, + "loss": 2.998800754547119, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.030249621719121933, + "learning_rate": 0.000125422220031917, + "loss": 3.0187838077545166, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.031052686274051666, + "learning_rate": 0.00012520860459779972, + "loss": 2.984062433242798, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.02961024083197117, + "learning_rate": 0.0001249951324069986, + "loss": 2.970818042755127, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.03020382858812809, + "learning_rate": 0.0001247818036315042, + "loss": 2.9913196563720703, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.031914591789245605, + "learning_rate": 0.00012456861844319155, + "loss": 2.993145704269409, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.03061537817120552, + "learning_rate": 0.00012435557701382012, + "loss": 3.0131571292877197, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.029960963875055313, + "learning_rate": 0.0001241426795150336, + "loss": 3.0017619132995605, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.030348235741257668, + "learning_rate": 0.00012392992611835973, + "loss": 3.0220556259155273, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.03030729480087757, + "learning_rate": 0.0001237173169952098, + "loss": 3.0499038696289062, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.030809229239821434, + "learning_rate": 0.00012350485231687954, + "loss": 2.986651659011841, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.029438305646181107, + "learning_rate": 0.0001232925322545476, + "loss": 3.0351314544677734, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.029887216165661812, + "learning_rate": 0.0001230803569792765, + "loss": 3.021660327911377, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.02955588512122631, + "learning_rate": 0.0001228683266620121, + "loss": 2.9990694522857666, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.029586469754576683, + "learning_rate": 0.00012265644147358326, + "loss": 3.0296952724456787, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.030632557347416878, + "learning_rate": 0.00012244470158470226, + "loss": 3.0124430656433105, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.030579863116145134, + "learning_rate": 0.00012223310716596387, + "loss": 3.0381762981414795, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.03282996267080307, + "learning_rate": 0.00012202165838784601, + "loss": 3.0329041481018066, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.031176412478089333, + "learning_rate": 0.00012181035542070913, + "loss": 3.037598133087158, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.028640294447541237, + "learning_rate": 0.00012159919843479617, + "loss": 3.0004701614379883, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.0326533280313015, + "learning_rate": 0.00012138818760023257, + "loss": 3.0328853130340576, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.03041139990091324, + "learning_rate": 0.00012117732308702592, + "loss": 3.0149893760681152, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.030366195365786552, + "learning_rate": 0.00012096660506506605, + "loss": 3.0379695892333984, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.03106893040239811, + "learning_rate": 0.00012075603370412443, + "loss": 3.0284576416015625, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.0288887657225132, + "learning_rate": 0.00012054560917385476, + "loss": 3.028569221496582, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.03154266998171806, + "learning_rate": 0.00012033533164379225, + "loss": 3.0737850666046143, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.0305064395070076, + "learning_rate": 0.0001201252012833537, + "loss": 3.0744194984436035, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.030967431142926216, + "learning_rate": 0.00011991521826183747, + "loss": 3.0170607566833496, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.036945272237062454, + "learning_rate": 0.00011970538274842282, + "loss": 3.0355610847473145, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.329805374145508, + "eval_runtime": 64.1339, + "eval_samples_per_second": 38.077, + "eval_steps_per_second": 1.201, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.03236053138971329, + "learning_rate": 0.00011949569491217073, + "loss": 2.9230728149414062, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.0395631417632103, + "learning_rate": 0.00011928615492202269, + "loss": 2.92578125, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.0361955352127552, + "learning_rate": 0.0001190767629468014, + "loss": 2.937826633453369, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.035918429493904114, + "learning_rate": 0.00011886751915521007, + "loss": 2.9481167793273926, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.03309318795800209, + "learning_rate": 0.00011865842371583278, + "loss": 2.8919506072998047, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0327315591275692, + "learning_rate": 0.00011844947679713396, + "loss": 2.903585433959961, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.033141329884529114, + "learning_rate": 0.00011824067856745812, + "loss": 2.942802667617798, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.033966805785894394, + "learning_rate": 0.00011803202919503064, + "loss": 2.9210848808288574, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.03218021243810654, + "learning_rate": 0.00011782352884795615, + "loss": 2.933588981628418, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.03272828459739685, + "learning_rate": 0.00011761517769421983, + "loss": 2.9202046394348145, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.030764421448111534, + "learning_rate": 0.00011740697590168635, + "loss": 2.916121006011963, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.031897611916065216, + "learning_rate": 0.00011719892363810018, + "loss": 2.957679033279419, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.03199796751141548, + "learning_rate": 0.00011699102107108539, + "loss": 2.9276621341705322, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.030233927071094513, + "learning_rate": 0.00011678326836814507, + "loss": 2.9006078243255615, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.03035062551498413, + "learning_rate": 0.00011657566569666198, + "loss": 2.920341730117798, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.03029404580593109, + "learning_rate": 0.00011636821322389777, + "loss": 2.937875747680664, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.029371997341513634, + "learning_rate": 0.00011616091111699333, + "loss": 2.9832730293273926, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.030163884162902832, + "learning_rate": 0.00011595375954296786, + "loss": 2.94118332862854, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.03072655200958252, + "learning_rate": 0.00011574675866871997, + "loss": 2.9632580280303955, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.02975962497293949, + "learning_rate": 0.00011553990866102651, + "loss": 2.9346470832824707, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.029814405366778374, + "learning_rate": 0.00011533320968654265, + "loss": 2.9451403617858887, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.030135225504636765, + "learning_rate": 0.00011512666191180204, + "loss": 2.944918155670166, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.029395751655101776, + "learning_rate": 0.00011492026550321653, + "loss": 2.925586223602295, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.0306396484375, + "learning_rate": 0.00011471402062707607, + "loss": 2.9487359523773193, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.03110465221107006, + "learning_rate": 0.00011450792744954827, + "loss": 2.963491678237915, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.03155078738927841, + "learning_rate": 0.00011430198613667861, + "loss": 2.9355576038360596, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.030103784054517746, + "learning_rate": 0.00011409619685439064, + "loss": 2.953045606613159, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.030827803537249565, + "learning_rate": 0.00011389055976848477, + "loss": 2.968574047088623, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.0310911163687706, + "learning_rate": 0.00011368507504463914, + "loss": 2.977902412414551, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.030260782688856125, + "learning_rate": 0.00011347974284840914, + "loss": 2.94873309135437, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.030654428526759148, + "learning_rate": 0.00011327456334522725, + "loss": 2.9282753467559814, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.02993479184806347, + "learning_rate": 0.00011306953670040275, + "loss": 2.97076416015625, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.030244970694184303, + "learning_rate": 0.00011286466307912184, + "loss": 2.960441827774048, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.030003050342202187, + "learning_rate": 0.00011265994264644783, + "loss": 2.9672164916992188, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.030057862401008606, + "learning_rate": 0.00011245537556732001, + "loss": 2.961918830871582, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.029379919171333313, + "learning_rate": 0.00011225096200655442, + "loss": 2.9918394088745117, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.03060895763337612, + "learning_rate": 0.00011204670212884342, + "loss": 2.94891619682312, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.030228130519390106, + "learning_rate": 0.00011184259609875557, + "loss": 2.9726617336273193, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.029833678156137466, + "learning_rate": 0.00011163864408073522, + "loss": 2.9494335651397705, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.030842263251543045, + "learning_rate": 0.00011143484623910293, + "loss": 2.9580483436584473, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.029707176610827446, + "learning_rate": 0.00011123120273805496, + "loss": 2.963684320449829, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.030966786667704582, + "learning_rate": 0.0001110277137416632, + "loss": 2.9710958003997803, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.03137032687664032, + "learning_rate": 0.00011082437941387512, + "loss": 3.003589391708374, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.03123190440237522, + "learning_rate": 0.00011062119991851322, + "loss": 2.9742860794067383, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.030608588829636574, + "learning_rate": 0.000110418175419276, + "loss": 2.97404408454895, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.030018644407391548, + "learning_rate": 0.00011021530607973626, + "loss": 2.972231864929199, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.03056846372783184, + "learning_rate": 0.00011001259206334235, + "loss": 2.9871716499328613, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.02888435684144497, + "learning_rate": 0.00010981003353341721, + "loss": 2.9774169921875, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.03173990175127983, + "learning_rate": 0.00010960763065315864, + "loss": 2.941751480102539, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.030047327280044556, + "learning_rate": 0.00010940538358563906, + "loss": 2.990516185760498, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.030729815363883972, + "learning_rate": 0.00010920329249380509, + "loss": 2.9745054244995117, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.030600983649492264, + "learning_rate": 0.00010900135754047799, + "loss": 2.9676642417907715, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.03208782523870468, + "learning_rate": 0.00010879957888835307, + "loss": 3.006941556930542, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.031162556260824203, + "learning_rate": 0.00010859795669999981, + "loss": 2.9740395545959473, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.031870320439338684, + "learning_rate": 0.00010839649113786137, + "loss": 2.9809296131134033, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.03184300288558006, + "learning_rate": 0.00010819518236425514, + "loss": 2.9781875610351562, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.029629364609718323, + "learning_rate": 0.00010799403054137197, + "loss": 2.9912891387939453, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.03322628512978554, + "learning_rate": 0.00010779303583127609, + "loss": 2.9957847595214844, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.0322413295507431, + "learning_rate": 0.00010759219839590535, + "loss": 2.9894957542419434, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.030753307044506073, + "learning_rate": 0.00010739151839707089, + "loss": 3.0320589542388916, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.03447900339961052, + "learning_rate": 0.00010719099599645706, + "loss": 3.0189106464385986, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.032016605138778687, + "learning_rate": 0.0001069906313556208, + "loss": 3.0162312984466553, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.030838409438729286, + "learning_rate": 0.00010679042463599263, + "loss": 3.000263214111328, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.034556787461042404, + "learning_rate": 0.00010659037599887545, + "loss": 2.996680736541748, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.03266173601150513, + "learning_rate": 0.0001063904856054446, + "loss": 2.990510940551758, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.030973197892308235, + "learning_rate": 0.00010619075361674836, + "loss": 2.9767255783081055, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.03442178666591644, + "learning_rate": 0.0001059911801937071, + "loss": 2.9941515922546387, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.03215927258133888, + "learning_rate": 0.00010579176549711365, + "loss": 3.006618022918701, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.030949831008911133, + "learning_rate": 0.00010559250968763265, + "loss": 3.004136085510254, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.03256693854928017, + "learning_rate": 0.00010539341292580086, + "loss": 3.0229580402374268, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.02986278384923935, + "learning_rate": 0.00010519447537202729, + "loss": 2.9580869674682617, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.03597614914178848, + "learning_rate": 0.00010499569718659201, + "loss": 2.995757579803467, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.359405517578125, + "eval_runtime": 68.1721, + "eval_samples_per_second": 35.821, + "eval_steps_per_second": 1.129, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.03432514891028404, + "learning_rate": 0.00010479707852964713, + "loss": 2.922243118286133, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.0345693901181221, + "learning_rate": 0.00010459861956121611, + "loss": 2.9114043712615967, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.0349399633705616, + "learning_rate": 0.00010440032044119383, + "loss": 2.8668057918548584, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.03161955252289772, + "learning_rate": 0.00010420218132934614, + "loss": 2.8996763229370117, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.032124556601047516, + "learning_rate": 0.00010400420238531023, + "loss": 2.8662075996398926, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.03291172534227371, + "learning_rate": 0.00010380638376859416, + "loss": 2.9088220596313477, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.033552397042512894, + "learning_rate": 0.00010360872563857682, + "loss": 2.9076967239379883, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.03218761831521988, + "learning_rate": 0.0001034112281545079, + "loss": 2.8726117610931396, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.03289451450109482, + "learning_rate": 0.00010321389147550725, + "loss": 2.9127345085144043, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.033019524067640305, + "learning_rate": 0.00010301671576056588, + "loss": 2.920897960662842, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.03153533488512039, + "learning_rate": 0.00010281970116854436, + "loss": 2.9124374389648438, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.032293178141117096, + "learning_rate": 0.00010262284785817392, + "loss": 2.877277374267578, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.031969670206308365, + "learning_rate": 0.00010242615598805574, + "loss": 2.921635150909424, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.03023488074541092, + "learning_rate": 0.00010222962571666088, + "loss": 2.9336061477661133, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.03215106576681137, + "learning_rate": 0.00010203325720233032, + "loss": 2.905877113342285, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.030415233224630356, + "learning_rate": 0.00010183705060327433, + "loss": 2.909331798553467, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.031150560826063156, + "learning_rate": 0.00010164100607757346, + "loss": 2.9302995204925537, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.03152180463075638, + "learning_rate": 0.00010144512378317687, + "loss": 2.9171547889709473, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.02996763028204441, + "learning_rate": 0.00010124940387790354, + "loss": 2.916459798812866, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.030544105917215347, + "learning_rate": 0.00010105384651944148, + "loss": 2.898505449295044, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.030963370576500893, + "learning_rate": 0.00010085845186534769, + "loss": 2.948704719543457, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.03191790357232094, + "learning_rate": 0.00010066322007304819, + "loss": 2.9182960987091064, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.02956460975110531, + "learning_rate": 0.00010046815129983757, + "loss": 2.9211082458496094, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.031227655708789825, + "learning_rate": 0.00010027324570287925, + "loss": 2.942269802093506, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.031859587877988815, + "learning_rate": 0.00010007850343920519, + "loss": 2.9169349670410156, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.03045276179909706, + "learning_rate": 9.988392466571572e-05, + "loss": 2.9240684509277344, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.031577471643686295, + "learning_rate": 9.968950953917945e-05, + "loss": 2.9135894775390625, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.031031791120767593, + "learning_rate": 9.949525821623309e-05, + "loss": 2.913355827331543, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.030012110248208046, + "learning_rate": 9.930117085338156e-05, + "loss": 2.8860886096954346, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.029772799462080002, + "learning_rate": 9.910724760699731e-05, + "loss": 2.974332809448242, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.029961293563246727, + "learning_rate": 9.891348863332093e-05, + "loss": 2.943694591522217, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.029537590220570564, + "learning_rate": 9.871989408846051e-05, + "loss": 2.93180513381958, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.030834559351205826, + "learning_rate": 9.852646412839178e-05, + "loss": 2.9442358016967773, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.02998523600399494, + "learning_rate": 9.833319890895756e-05, + "loss": 2.9336910247802734, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.030030962079763412, + "learning_rate": 9.814009858586817e-05, + "loss": 2.93593692779541, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.030250679701566696, + "learning_rate": 9.794716331470131e-05, + "loss": 2.9370553493499756, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.02942339889705181, + "learning_rate": 9.77543932509012e-05, + "loss": 2.9406495094299316, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.030448907986283302, + "learning_rate": 9.756178854977925e-05, + "loss": 2.9552698135375977, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.03078317642211914, + "learning_rate": 9.736934936651362e-05, + "loss": 2.933631420135498, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.029188675805926323, + "learning_rate": 9.717707585614916e-05, + "loss": 2.944002628326416, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.03153637796640396, + "learning_rate": 9.698496817359698e-05, + "loss": 2.9708166122436523, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.030299564823508263, + "learning_rate": 9.679302647363476e-05, + "loss": 2.9570717811584473, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.03169458732008934, + "learning_rate": 9.660125091090675e-05, + "loss": 2.942774772644043, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.031527288258075714, + "learning_rate": 9.64096416399228e-05, + "loss": 2.9711742401123047, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.0314360186457634, + "learning_rate": 9.621819881505918e-05, + "loss": 2.9509239196777344, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.032102152705192566, + "learning_rate": 9.602692259055767e-05, + "loss": 2.9443840980529785, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.03111223131418228, + "learning_rate": 9.583581312052646e-05, + "loss": 2.972594738006592, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.0322272814810276, + "learning_rate": 9.564487055893867e-05, + "loss": 2.9361982345581055, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.033640168607234955, + "learning_rate": 9.545409505963338e-05, + "loss": 2.9726009368896484, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.02979348786175251, + "learning_rate": 9.526348677631499e-05, + "loss": 2.9978041648864746, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.030464721843600273, + "learning_rate": 9.507304586255311e-05, + "loss": 2.949810028076172, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.03170578181743622, + "learning_rate": 9.488277247178267e-05, + "loss": 2.9677181243896484, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.02996704913675785, + "learning_rate": 9.469266675730319e-05, + "loss": 2.974087953567505, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.030277637764811516, + "learning_rate": 9.450272887227983e-05, + "loss": 2.95644211769104, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.030859284102916718, + "learning_rate": 9.431295896974182e-05, + "loss": 2.957258701324463, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.029737796634435654, + "learning_rate": 9.412335720258341e-05, + "loss": 2.9502854347229004, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.029822375625371933, + "learning_rate": 9.393392372356335e-05, + "loss": 2.9625749588012695, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.030455606058239937, + "learning_rate": 9.374465868530477e-05, + "loss": 2.9669153690338135, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.029491793364286423, + "learning_rate": 9.355556224029515e-05, + "loss": 2.952991247177124, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.030326122418045998, + "learning_rate": 9.336663454088593e-05, + "loss": 2.978355646133423, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.03011198714375496, + "learning_rate": 9.317787573929282e-05, + "loss": 2.976996421813965, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.029053322970867157, + "learning_rate": 9.298928598759541e-05, + "loss": 2.9754371643066406, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.02925761416554451, + "learning_rate": 9.280086543773698e-05, + "loss": 2.94271183013916, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.029678862541913986, + "learning_rate": 9.26126142415247e-05, + "loss": 2.9386143684387207, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.031623054295778275, + "learning_rate": 9.2424532550629e-05, + "loss": 2.9660329818725586, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.030133802443742752, + "learning_rate": 9.223662051658408e-05, + "loss": 2.9580249786376953, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.029482144862413406, + "learning_rate": 9.204887829078709e-05, + "loss": 2.944474697113037, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.030708568170666695, + "learning_rate": 9.186130602449861e-05, + "loss": 2.967862367630005, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.028793595731258392, + "learning_rate": 9.167390386884224e-05, + "loss": 2.9581642150878906, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.03083081543445587, + "learning_rate": 9.148667197480455e-05, + "loss": 2.957261323928833, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.030180038884282112, + "learning_rate": 9.129961049323494e-05, + "loss": 2.9579484462738037, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.034881673753261566, + "learning_rate": 9.111271957484519e-05, + "loss": 2.9007070064544678, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.38953161239624, + "eval_runtime": 71.4964, + "eval_samples_per_second": 34.156, + "eval_steps_per_second": 1.077, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.0335528701543808, + "learning_rate": 9.092599937021031e-05, + "loss": 2.8542537689208984, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.03163984417915344, + "learning_rate": 9.073945002976715e-05, + "loss": 2.864593505859375, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.034005917608737946, + "learning_rate": 9.055307170381522e-05, + "loss": 2.855072498321533, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.03235425800085068, + "learning_rate": 9.036686454251613e-05, + "loss": 2.8504371643066406, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.03171464428305626, + "learning_rate": 9.018082869589369e-05, + "loss": 2.8924291133880615, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.0333196297287941, + "learning_rate": 8.999496431383368e-05, + "loss": 2.8764119148254395, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.03050111047923565, + "learning_rate": 8.980927154608341e-05, + "loss": 2.874119281768799, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.032114267349243164, + "learning_rate": 8.962375054225255e-05, + "loss": 2.9059693813323975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.03141588345170021, + "learning_rate": 8.943840145181177e-05, + "loss": 2.879429340362549, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.03201865404844284, + "learning_rate": 8.925322442409361e-05, + "loss": 2.898519277572632, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.03152940049767494, + "learning_rate": 8.906821960829163e-05, + "loss": 2.866131544113159, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.032520636916160583, + "learning_rate": 8.888338715346113e-05, + "loss": 2.888209819793701, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.03244909271597862, + "learning_rate": 8.869872720851831e-05, + "loss": 2.888913631439209, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.03160687908530235, + "learning_rate": 8.851423992224012e-05, + "loss": 2.905120372772217, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.031518299132585526, + "learning_rate": 8.832992544326473e-05, + "loss": 2.8985815048217773, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.0304370429366827, + "learning_rate": 8.814578392009104e-05, + "loss": 2.8531291484832764, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.03239843621850014, + "learning_rate": 8.796181550107857e-05, + "loss": 2.876098394393921, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.03194732964038849, + "learning_rate": 8.777802033444712e-05, + "loss": 2.908271074295044, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.03145617991685867, + "learning_rate": 8.75943985682774e-05, + "loss": 2.874128580093384, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.031182898208498955, + "learning_rate": 8.741095035051017e-05, + "loss": 2.894904613494873, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.0323818176984787, + "learning_rate": 8.722767582894613e-05, + "loss": 2.9005861282348633, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.030969973653554916, + "learning_rate": 8.704457515124636e-05, + "loss": 2.9236338138580322, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.03238993510603905, + "learning_rate": 8.686164846493176e-05, + "loss": 2.8811020851135254, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.030876649543642998, + "learning_rate": 8.667889591738317e-05, + "loss": 2.8802058696746826, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.032251305878162384, + "learning_rate": 8.649631765584083e-05, + "loss": 2.882112503051758, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.03334304690361023, + "learning_rate": 8.631391382740482e-05, + "loss": 2.9290661811828613, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.03299225866794586, + "learning_rate": 8.613168457903459e-05, + "loss": 2.908268928527832, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.03335456922650337, + "learning_rate": 8.594963005754901e-05, + "loss": 2.9156572818756104, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.030786098912358284, + "learning_rate": 8.57677504096261e-05, + "loss": 2.8981709480285645, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.03379470854997635, + "learning_rate": 8.558604578180301e-05, + "loss": 2.9483864307403564, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.03020016849040985, + "learning_rate": 8.540451632047593e-05, + "loss": 2.908958911895752, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.032774876803159714, + "learning_rate": 8.522316217189972e-05, + "loss": 2.899505138397217, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.030589981004595757, + "learning_rate": 8.504198348218821e-05, + "loss": 2.9257779121398926, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.03366618603467941, + "learning_rate": 8.486098039731384e-05, + "loss": 2.8749284744262695, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.03038695827126503, + "learning_rate": 8.46801530631075e-05, + "loss": 2.9127252101898193, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.03400666266679764, + "learning_rate": 8.449950162525859e-05, + "loss": 2.922375202178955, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.031702589243650436, + "learning_rate": 8.431902622931443e-05, + "loss": 2.8886446952819824, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.03223840892314911, + "learning_rate": 8.413872702068119e-05, + "loss": 2.906278610229492, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.0327419750392437, + "learning_rate": 8.395860414462238e-05, + "loss": 2.9436042308807373, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.030698910355567932, + "learning_rate": 8.377865774625985e-05, + "loss": 2.897615909576416, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.03250030800700188, + "learning_rate": 8.359888797057316e-05, + "loss": 2.928028106689453, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.030623450875282288, + "learning_rate": 8.341929496239958e-05, + "loss": 2.9268975257873535, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.030981825664639473, + "learning_rate": 8.323987886643402e-05, + "loss": 2.938185691833496, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.03139727562665939, + "learning_rate": 8.306063982722855e-05, + "loss": 2.9397618770599365, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.029906271025538445, + "learning_rate": 8.288157798919318e-05, + "loss": 2.921111583709717, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.03110392577946186, + "learning_rate": 8.27026934965945e-05, + "loss": 2.899641513824463, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.03060143068432808, + "learning_rate": 8.25239864935566e-05, + "loss": 2.922635078430176, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.03061371110379696, + "learning_rate": 8.23454571240605e-05, + "loss": 2.925189733505249, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.03080526366829872, + "learning_rate": 8.216710553194416e-05, + "loss": 2.90329647064209, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.030412547290325165, + "learning_rate": 8.198893186090222e-05, + "loss": 2.9444844722747803, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.03024943359196186, + "learning_rate": 8.181093625448585e-05, + "loss": 2.9165825843811035, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.030653616413474083, + "learning_rate": 8.163311885610299e-05, + "loss": 2.9171371459960938, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.03027702495455742, + "learning_rate": 8.145547980901791e-05, + "loss": 2.9399702548980713, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.030466679483652115, + "learning_rate": 8.127801925635126e-05, + "loss": 2.888066291809082, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.02997267246246338, + "learning_rate": 8.110073734107954e-05, + "loss": 2.906545877456665, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.03026283159852028, + "learning_rate": 8.092363420603584e-05, + "loss": 2.9359335899353027, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.0293166171759367, + "learning_rate": 8.074670999390896e-05, + "loss": 2.9118804931640625, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.031117571517825127, + "learning_rate": 8.056996484724342e-05, + "loss": 2.968198299407959, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.03041299246251583, + "learning_rate": 8.039339890843958e-05, + "loss": 2.9333484172821045, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.03037327155470848, + "learning_rate": 8.02170123197535e-05, + "loss": 2.895859956741333, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.03161945194005966, + "learning_rate": 8.004080522329674e-05, + "loss": 2.9491944313049316, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.031058384105563164, + "learning_rate": 7.986477776103601e-05, + "loss": 2.926038980484009, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.030037537217140198, + "learning_rate": 7.968893007479343e-05, + "loss": 2.941394805908203, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.030748063698410988, + "learning_rate": 7.951326230624658e-05, + "loss": 2.998904228210449, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.030324939638376236, + "learning_rate": 7.933777459692756e-05, + "loss": 2.9226560592651367, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.030745431780815125, + "learning_rate": 7.916246708822373e-05, + "loss": 2.977520227432251, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.02952338382601738, + "learning_rate": 7.898733992137715e-05, + "loss": 2.9365828037261963, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.030132673680782318, + "learning_rate": 7.881239323748475e-05, + "loss": 2.9574036598205566, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.030379803851246834, + "learning_rate": 7.863762717749771e-05, + "loss": 2.9329113960266113, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.029670827090740204, + "learning_rate": 7.846304188222189e-05, + "loss": 2.951117515563965, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.029823293909430504, + "learning_rate": 7.828863749231777e-05, + "loss": 2.9586987495422363, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.03546741604804993, + "learning_rate": 7.811441414829958e-05, + "loss": 2.9378662109375, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.417913436889648, + "eval_runtime": 69.6157, + "eval_samples_per_second": 35.078, + "eval_steps_per_second": 1.106, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.033859945833683014, + "learning_rate": 7.794037199053598e-05, + "loss": 2.8493661880493164, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.032931916415691376, + "learning_rate": 7.776651115924959e-05, + "loss": 2.844602108001709, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.03232948109507561, + "learning_rate": 7.759283179451704e-05, + "loss": 2.8463873863220215, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.032815027981996536, + "learning_rate": 7.741933403626848e-05, + "loss": 2.846658706665039, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.03269147500395775, + "learning_rate": 7.724601802428805e-05, + "loss": 2.8286149501800537, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.03337012976408005, + "learning_rate": 7.707288389821329e-05, + "loss": 2.9202351570129395, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.031733594834804535, + "learning_rate": 7.689993179753519e-05, + "loss": 2.8601999282836914, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.032877493649721146, + "learning_rate": 7.672716186159829e-05, + "loss": 2.8566389083862305, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.033852867782115936, + "learning_rate": 7.655457422959993e-05, + "loss": 2.859797477722168, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.03143248334527016, + "learning_rate": 7.638216904059122e-05, + "loss": 2.8411474227905273, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.033198583871126175, + "learning_rate": 7.620994643347559e-05, + "loss": 2.888702392578125, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.029946891590952873, + "learning_rate": 7.60379065470098e-05, + "loss": 2.8332901000976562, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.031660862267017365, + "learning_rate": 7.586604951980326e-05, + "loss": 2.8255481719970703, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.030189910903573036, + "learning_rate": 7.569437549031813e-05, + "loss": 2.8664190769195557, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.03273668885231018, + "learning_rate": 7.55228845968691e-05, + "loss": 2.886531114578247, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.030584124848246574, + "learning_rate": 7.535157697762313e-05, + "loss": 2.872990131378174, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.032543931156396866, + "learning_rate": 7.518045277059979e-05, + "loss": 2.891528606414795, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.031016048043966293, + "learning_rate": 7.500951211367068e-05, + "loss": 2.8759872913360596, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.03190434351563454, + "learning_rate": 7.483875514455979e-05, + "loss": 2.8757526874542236, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.030996622517704964, + "learning_rate": 7.466818200084264e-05, + "loss": 2.862654447555542, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.03289523348212242, + "learning_rate": 7.449779281994712e-05, + "loss": 2.9064857959747314, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.03322640061378479, + "learning_rate": 7.43275877391528e-05, + "loss": 2.8921048641204834, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.0303682629019022, + "learning_rate": 7.415756689559061e-05, + "loss": 2.886458396911621, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.033412594348192215, + "learning_rate": 7.398773042624341e-05, + "loss": 2.8638253211975098, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.030457086861133575, + "learning_rate": 7.381807846794532e-05, + "loss": 2.835578441619873, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.03225106745958328, + "learning_rate": 7.3648611157382e-05, + "loss": 2.8652632236480713, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.03178466483950615, + "learning_rate": 7.34793286310899e-05, + "loss": 2.8796353340148926, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.03075973503291607, + "learning_rate": 7.331023102545716e-05, + "loss": 2.895150661468506, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.02984696812927723, + "learning_rate": 7.314131847672268e-05, + "loss": 2.882878303527832, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.030945835635066032, + "learning_rate": 7.297259112097608e-05, + "loss": 2.898838520050049, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.030904723331332207, + "learning_rate": 7.280404909415801e-05, + "loss": 2.885678291320801, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.030709490180015564, + "learning_rate": 7.263569253205973e-05, + "loss": 2.8645198345184326, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.030528701841831207, + "learning_rate": 7.24675215703232e-05, + "loss": 2.8877408504486084, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.030471473932266235, + "learning_rate": 7.229953634444055e-05, + "loss": 2.9136905670166016, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.030309896916151047, + "learning_rate": 7.213173698975444e-05, + "loss": 2.9238810539245605, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.030605493113398552, + "learning_rate": 7.1964123641458e-05, + "loss": 2.8624460697174072, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.029686832800507545, + "learning_rate": 7.179669643459405e-05, + "loss": 2.9085543155670166, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.030039243400096893, + "learning_rate": 7.162945550405575e-05, + "loss": 2.8624000549316406, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.02953776717185974, + "learning_rate": 7.146240098458613e-05, + "loss": 2.8893938064575195, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.029651308432221413, + "learning_rate": 7.129553301077801e-05, + "loss": 2.914975166320801, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.030902806669473648, + "learning_rate": 7.112885171707378e-05, + "loss": 2.887071132659912, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.02948773466050625, + "learning_rate": 7.096235723776562e-05, + "loss": 2.8893041610717773, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.03049607202410698, + "learning_rate": 7.079604970699514e-05, + "loss": 2.902736186981201, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.030077926814556122, + "learning_rate": 7.062992925875332e-05, + "loss": 2.8936967849731445, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.0303986556828022, + "learning_rate": 7.046399602688045e-05, + "loss": 2.9223172664642334, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.030650658532977104, + "learning_rate": 7.029825014506572e-05, + "loss": 2.9140920639038086, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.029562659561634064, + "learning_rate": 7.013269174684795e-05, + "loss": 2.918179988861084, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.03063463792204857, + "learning_rate": 6.996732096561422e-05, + "loss": 2.9222826957702637, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.030284354463219643, + "learning_rate": 6.980213793460092e-05, + "loss": 2.8685007095336914, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.029791876673698425, + "learning_rate": 6.963714278689304e-05, + "loss": 2.914102554321289, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.030152756720781326, + "learning_rate": 6.947233565542417e-05, + "loss": 2.8890669345855713, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.02957645244896412, + "learning_rate": 6.930771667297653e-05, + "loss": 2.8851962089538574, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.02963421493768692, + "learning_rate": 6.914328597218042e-05, + "loss": 2.9012060165405273, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.029279837384819984, + "learning_rate": 6.8979043685515e-05, + "loss": 2.949984073638916, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.0302849430590868, + "learning_rate": 6.881498994530708e-05, + "loss": 2.9049415588378906, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.029432371258735657, + "learning_rate": 6.865112488373186e-05, + "loss": 2.8919944763183594, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.030542058870196342, + "learning_rate": 6.84874486328125e-05, + "loss": 2.9168801307678223, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.02911970019340515, + "learning_rate": 6.832396132441993e-05, + "loss": 2.9342129230499268, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.030335791409015656, + "learning_rate": 6.816066309027309e-05, + "loss": 2.918611764907837, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.03031272068619728, + "learning_rate": 6.799755406193815e-05, + "loss": 2.8769750595092773, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.030200788751244545, + "learning_rate": 6.783463437082929e-05, + "loss": 2.8994460105895996, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.03032573312520981, + "learning_rate": 6.767190414820792e-05, + "loss": 2.920720100402832, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.03180326521396637, + "learning_rate": 6.750936352518284e-05, + "loss": 2.891145706176758, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.029451677575707436, + "learning_rate": 6.734701263271011e-05, + "loss": 2.9057083129882812, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.030261319130659103, + "learning_rate": 6.718485160159289e-05, + "loss": 2.9132461547851562, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.031257566064596176, + "learning_rate": 6.702288056248145e-05, + "loss": 2.9089136123657227, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.03043205663561821, + "learning_rate": 6.686109964587285e-05, + "loss": 2.912537097930908, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.030708536505699158, + "learning_rate": 6.669950898211106e-05, + "loss": 2.8841137886047363, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.03044111281633377, + "learning_rate": 6.653810870138676e-05, + "loss": 2.921257257461548, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.029576802626252174, + "learning_rate": 6.637689893373729e-05, + "loss": 2.908468723297119, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.03031427040696144, + "learning_rate": 6.621587980904651e-05, + "loss": 2.936445713043213, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.036312974989414215, + "learning_rate": 6.605505145704437e-05, + "loss": 2.8997855186462402, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.443755149841309, + "eval_runtime": 62.9898, + "eval_samples_per_second": 38.768, + "eval_steps_per_second": 1.222, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.03337511047720909, + "learning_rate": 6.58944140073077e-05, + "loss": 2.8584513664245605, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.031415343284606934, + "learning_rate": 6.573396758925891e-05, + "loss": 2.8529202938079834, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.033063437789678574, + "learning_rate": 6.557371233216692e-05, + "loss": 2.814058780670166, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.03153088316321373, + "learning_rate": 6.541364836514646e-05, + "loss": 2.80460786819458, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.03227711841464043, + "learning_rate": 6.525377581715829e-05, + "loss": 2.858191967010498, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.030423127114772797, + "learning_rate": 6.509409481700864e-05, + "loss": 2.845815420150757, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.032030411064624786, + "learning_rate": 6.49346054933496e-05, + "loss": 2.837918281555176, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.030464086681604385, + "learning_rate": 6.477530797467911e-05, + "loss": 2.841099739074707, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.03194798156619072, + "learning_rate": 6.461620238934006e-05, + "loss": 2.843568801879883, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.03152863308787346, + "learning_rate": 6.445728886552109e-05, + "loss": 2.8779497146606445, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.03201183304190636, + "learning_rate": 6.429856753125573e-05, + "loss": 2.906585216522217, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.02999979443848133, + "learning_rate": 6.414003851442318e-05, + "loss": 2.84859561920166, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.03106156550347805, + "learning_rate": 6.398170194274722e-05, + "loss": 2.814643383026123, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.03118651732802391, + "learning_rate": 6.38235579437968e-05, + "loss": 2.8370001316070557, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.031022636219859123, + "learning_rate": 6.366560664498572e-05, + "loss": 2.8466384410858154, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.03163835406303406, + "learning_rate": 6.350784817357242e-05, + "loss": 2.808440685272217, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.03071477636694908, + "learning_rate": 6.335028265666023e-05, + "loss": 2.850621223449707, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.03041849471628666, + "learning_rate": 6.319291022119653e-05, + "loss": 2.818002939224243, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.03104640357196331, + "learning_rate": 6.303573099397378e-05, + "loss": 2.8375182151794434, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.02970169298350811, + "learning_rate": 6.287874510162821e-05, + "loss": 2.871188163757324, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.03146181255578995, + "learning_rate": 6.272195267064063e-05, + "loss": 2.8356661796569824, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.030204759910702705, + "learning_rate": 6.25653538273358e-05, + "loss": 2.861675262451172, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.029957789927721024, + "learning_rate": 6.240894869788267e-05, + "loss": 2.8493146896362305, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.031224241480231285, + "learning_rate": 6.225273740829404e-05, + "loss": 2.876760959625244, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 0.02914278209209442, + "learning_rate": 6.209672008442635e-05, + "loss": 2.838275909423828, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.03204638510942459, + "learning_rate": 6.19408968519801e-05, + "loss": 2.8679757118225098, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.030257491394877434, + "learning_rate": 6.178526783649916e-05, + "loss": 2.8245105743408203, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.031307291239500046, + "learning_rate": 6.162983316337109e-05, + "loss": 2.852024793624878, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 0.030594198033213615, + "learning_rate": 6.147459295782676e-05, + "loss": 2.883610248565674, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.03119277022778988, + "learning_rate": 6.131954734494045e-05, + "loss": 2.866126775741577, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.0302574560046196, + "learning_rate": 6.11646964496296e-05, + "loss": 2.857886791229248, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.029702702537178993, + "learning_rate": 6.101004039665471e-05, + "loss": 2.8324077129364014, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.029572894796729088, + "learning_rate": 6.085557931061937e-05, + "loss": 2.878951072692871, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.03024783544242382, + "learning_rate": 6.070131331597015e-05, + "loss": 2.900284767150879, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.02952091582119465, + "learning_rate": 6.054724253699636e-05, + "loss": 2.8402292728424072, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.03006756864488125, + "learning_rate": 6.03933670978301e-05, + "loss": 2.8609299659729004, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.029146816581487656, + "learning_rate": 6.0239687122445826e-05, + "loss": 2.887960433959961, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.03026733547449112, + "learning_rate": 6.0086202734661e-05, + "loss": 2.852726459503174, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.029839467257261276, + "learning_rate": 5.9932914058135004e-05, + "loss": 2.8559489250183105, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.029919864609837532, + "learning_rate": 5.977982121636984e-05, + "loss": 2.880255937576294, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.03074759431183338, + "learning_rate": 5.962692433270962e-05, + "loss": 2.8586673736572266, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.029141079634428024, + "learning_rate": 5.9474223530340634e-05, + "loss": 2.8813042640686035, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.031023059040308, + "learning_rate": 5.932171893229124e-05, + "loss": 2.8355565071105957, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.029786985367536545, + "learning_rate": 5.916941066143137e-05, + "loss": 2.872175693511963, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.030190780758857727, + "learning_rate": 5.9017298840473364e-05, + "loss": 2.874236583709717, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.029669199138879776, + "learning_rate": 5.8865383591970775e-05, + "loss": 2.8723950386047363, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.030073825269937515, + "learning_rate": 5.8713665038319e-05, + "loss": 2.863720417022705, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.02976175583899021, + "learning_rate": 5.856214330175498e-05, + "loss": 2.883267402648926, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.030071711167693138, + "learning_rate": 5.841081850435704e-05, + "loss": 2.8690664768218994, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.031068995594978333, + "learning_rate": 5.825969076804488e-05, + "loss": 2.9107182025909424, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.030393972992897034, + "learning_rate": 5.810876021457925e-05, + "loss": 2.859959602355957, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.030435480177402496, + "learning_rate": 5.795802696556222e-05, + "loss": 2.9003896713256836, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.02982497587800026, + "learning_rate": 5.78074911424369e-05, + "loss": 2.857231378555298, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.030999863520264626, + "learning_rate": 5.765715286648734e-05, + "loss": 2.8567168712615967, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.029698362573981285, + "learning_rate": 5.7507012258838154e-05, + "loss": 2.8949484825134277, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.031584806740283966, + "learning_rate": 5.7357069440455095e-05, + "loss": 2.8703250885009766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.02952221967279911, + "learning_rate": 5.7207324532144454e-05, + "loss": 2.8914244174957275, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.03138599172234535, + "learning_rate": 5.7057777654552785e-05, + "loss": 2.917753219604492, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.029678434133529663, + "learning_rate": 5.690842892816741e-05, + "loss": 2.8854236602783203, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.03174183890223503, + "learning_rate": 5.675927847331593e-05, + "loss": 2.8508100509643555, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.03006329946219921, + "learning_rate": 5.661032641016619e-05, + "loss": 2.883819103240967, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.031188689172267914, + "learning_rate": 5.6461572858726086e-05, + "loss": 2.932220458984375, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.029729023575782776, + "learning_rate": 5.6313017938843595e-05, + "loss": 2.894927978515625, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.02953082136809826, + "learning_rate": 5.6164661770207026e-05, + "loss": 2.896371364593506, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.030423447489738464, + "learning_rate": 5.6016504472344e-05, + "loss": 2.9003541469573975, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.029901325702667236, + "learning_rate": 5.5868546164622284e-05, + "loss": 2.8726208209991455, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.030071355402469635, + "learning_rate": 5.572078696624917e-05, + "loss": 2.881291151046753, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.029847780242562294, + "learning_rate": 5.557322699627168e-05, + "loss": 2.944478988647461, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.029794953763484955, + "learning_rate": 5.542586637357607e-05, + "loss": 2.8625364303588867, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.03029358573257923, + "learning_rate": 5.527870521688817e-05, + "loss": 2.9106764793395996, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.031095417216420174, + "learning_rate": 5.513174364477311e-05, + "loss": 2.8879761695861816, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.03671329468488693, + "learning_rate": 5.498498177563518e-05, + "loss": 2.877009391784668, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 4.470407962799072, + "eval_runtime": 57.8161, + "eval_samples_per_second": 42.237, + "eval_steps_per_second": 1.332, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.03383656591176987, + "learning_rate": 5.4838419727717716e-05, + "loss": 2.8301892280578613, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.030587317422032356, + "learning_rate": 5.4692057619103144e-05, + "loss": 2.809070587158203, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.034895412623882294, + "learning_rate": 5.454589556771286e-05, + "loss": 2.845365285873413, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.03129304572939873, + "learning_rate": 5.439993369130681e-05, + "loss": 2.8176681995391846, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.03369338437914848, + "learning_rate": 5.4254172107483924e-05, + "loss": 2.820741653442383, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.030895931646227837, + "learning_rate": 5.410861093368169e-05, + "loss": 2.810272693634033, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.03239824250340462, + "learning_rate": 5.3963250287176094e-05, + "loss": 2.855811595916748, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.03123638592660427, + "learning_rate": 5.381809028508168e-05, + "loss": 2.7941160202026367, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.031117001548409462, + "learning_rate": 5.367313104435104e-05, + "loss": 2.784010887145996, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.03162471950054169, + "learning_rate": 5.352837268177544e-05, + "loss": 2.8337666988372803, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.0316263884305954, + "learning_rate": 5.3383815313983916e-05, + "loss": 2.8019790649414062, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.031872715801000595, + "learning_rate": 5.323945905744379e-05, + "loss": 2.8156003952026367, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.030252376571297646, + "learning_rate": 5.3095304028460284e-05, + "loss": 2.7971153259277344, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.03168163076043129, + "learning_rate": 5.2951350343176495e-05, + "loss": 2.791532039642334, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.03014831617474556, + "learning_rate": 5.2807598117573384e-05, + "loss": 2.8359432220458984, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.03142879158258438, + "learning_rate": 5.266404746746939e-05, + "loss": 2.846064567565918, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.030883461236953735, + "learning_rate": 5.252069850852072e-05, + "loss": 2.8668103218078613, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.03189866617321968, + "learning_rate": 5.237755135622107e-05, + "loss": 2.8214638233184814, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.029116913676261902, + "learning_rate": 5.223460612590159e-05, + "loss": 2.850940704345703, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.030518673360347748, + "learning_rate": 5.209186293273044e-05, + "loss": 2.844576358795166, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.029227688908576965, + "learning_rate": 5.1949321891713446e-05, + "loss": 2.797168493270874, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.029839880764484406, + "learning_rate": 5.180698311769338e-05, + "loss": 2.8534903526306152, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.03011847473680973, + "learning_rate": 5.166484672534983e-05, + "loss": 2.8278157711029053, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.03027244098484516, + "learning_rate": 5.152291282919961e-05, + "loss": 2.858158588409424, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.030171507969498634, + "learning_rate": 5.138118154359632e-05, + "loss": 2.8523292541503906, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.029657812789082527, + "learning_rate": 5.1239652982730335e-05, + "loss": 2.8015360832214355, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.028928041458129883, + "learning_rate": 5.1098327260628466e-05, + "loss": 2.8357324600219727, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.029674772173166275, + "learning_rate": 5.09572044911545e-05, + "loss": 2.8269171714782715, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.03013371117413044, + "learning_rate": 5.0816284788008527e-05, + "loss": 2.822054862976074, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.02960435301065445, + "learning_rate": 5.067556826472684e-05, + "loss": 2.8452866077423096, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.030388740822672844, + "learning_rate": 5.053505503468228e-05, + "loss": 2.855799674987793, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.029823970049619675, + "learning_rate": 5.0394745211083785e-05, + "loss": 2.8262312412261963, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.030613111332058907, + "learning_rate": 5.025463890697655e-05, + "loss": 2.8490824699401855, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.029467277228832245, + "learning_rate": 5.011473623524159e-05, + "loss": 2.855149984359741, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.030264565721154213, + "learning_rate": 4.9975037308595864e-05, + "loss": 2.849930763244629, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.02961486577987671, + "learning_rate": 4.983554223959257e-05, + "loss": 2.843088150024414, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.029051944613456726, + "learning_rate": 4.9696251140620134e-05, + "loss": 2.825998067855835, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.03038879856467247, + "learning_rate": 4.9557164123902924e-05, + "loss": 2.8565316200256348, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.02917695790529251, + "learning_rate": 4.9418281301500844e-05, + "loss": 2.8733534812927246, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.03011898323893547, + "learning_rate": 4.9279602785309365e-05, + "loss": 2.8347158432006836, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.03015534020960331, + "learning_rate": 4.914112868705911e-05, + "loss": 2.878455400466919, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.02992299199104309, + "learning_rate": 4.900285911831624e-05, + "loss": 2.861504554748535, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.030217986553907394, + "learning_rate": 4.886479419048201e-05, + "loss": 2.8451390266418457, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.029564134776592255, + "learning_rate": 4.872693401479292e-05, + "loss": 2.8065595626831055, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.03130771592259407, + "learning_rate": 4.858927870232039e-05, + "loss": 2.8506879806518555, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.030411556363105774, + "learning_rate": 4.845182836397068e-05, + "loss": 2.857304573059082, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.030868662521243095, + "learning_rate": 4.831458311048532e-05, + "loss": 2.8545470237731934, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.03002075105905533, + "learning_rate": 4.817754305244008e-05, + "loss": 2.889376640319824, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.030850842595100403, + "learning_rate": 4.804070830024577e-05, + "loss": 2.8634769916534424, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.03078618459403515, + "learning_rate": 4.7904078964147654e-05, + "loss": 2.8487653732299805, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.031029092147946358, + "learning_rate": 4.776765515422557e-05, + "loss": 2.897327184677124, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.03101371042430401, + "learning_rate": 4.763143698039371e-05, + "loss": 2.8642020225524902, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.03044370748102665, + "learning_rate": 4.749542455240043e-05, + "loss": 2.87313175201416, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.030529852956533432, + "learning_rate": 4.735961797982876e-05, + "loss": 2.849147081375122, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.029675021767616272, + "learning_rate": 4.7224017372095414e-05, + "loss": 2.8475494384765625, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.03002132847905159, + "learning_rate": 4.708862283845143e-05, + "loss": 2.8713879585266113, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.03048008307814598, + "learning_rate": 4.695343448798163e-05, + "loss": 2.8976492881774902, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.02930760383605957, + "learning_rate": 4.681845242960492e-05, + "loss": 2.850775718688965, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.029840776696801186, + "learning_rate": 4.668367677207398e-05, + "loss": 2.8743128776550293, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.030162667855620384, + "learning_rate": 4.654910762397499e-05, + "loss": 2.850478410720825, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.029736310243606567, + "learning_rate": 4.6414745093727913e-05, + "loss": 2.851759195327759, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.029385298490524292, + "learning_rate": 4.6280589289586255e-05, + "loss": 2.8473241329193115, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.028572317212820053, + "learning_rate": 4.614664031963692e-05, + "loss": 2.852155923843384, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.0300681721419096, + "learning_rate": 4.601289829180004e-05, + "loss": 2.871021270751953, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.0292969960719347, + "learning_rate": 4.587936331382934e-05, + "loss": 2.8470516204833984, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.029493169859051704, + "learning_rate": 4.574603549331151e-05, + "loss": 2.892655372619629, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.029351307079195976, + "learning_rate": 4.561291493766625e-05, + "loss": 2.8886585235595703, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.029827935621142387, + "learning_rate": 4.5480001754146455e-05, + "loss": 2.8802051544189453, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.029231170192360878, + "learning_rate": 4.5347296049837875e-05, + "loss": 2.876192092895508, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.029604196548461914, + "learning_rate": 4.52147979316592e-05, + "loss": 2.8420169353485107, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.030229991301894188, + "learning_rate": 4.5082507506361574e-05, + "loss": 2.86871075630188, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.0357736200094223, + "learning_rate": 4.495042488052904e-05, + "loss": 2.852806568145752, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 4.496274948120117, + "eval_runtime": 58.0976, + "eval_samples_per_second": 42.033, + "eval_steps_per_second": 1.325, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.03180048242211342, + "learning_rate": 4.481855016057839e-05, + "loss": 2.85683012008667, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.030023347586393356, + "learning_rate": 4.468688345275848e-05, + "loss": 2.814507484436035, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.03346574679017067, + "learning_rate": 4.455542486315086e-05, + "loss": 2.8209877014160156, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.030521582812070847, + "learning_rate": 4.442417449766938e-05, + "loss": 2.7859511375427246, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.033295996487140656, + "learning_rate": 4.429313246206014e-05, + "loss": 2.7879018783569336, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.029781939461827278, + "learning_rate": 4.416229886190119e-05, + "loss": 2.801133155822754, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.033207762986421585, + "learning_rate": 4.403167380260281e-05, + "loss": 2.8046321868896484, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.03027796931564808, + "learning_rate": 4.3901257389407475e-05, + "loss": 2.8235392570495605, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.03096267767250538, + "learning_rate": 4.3771049727389075e-05, + "loss": 2.8181068897247314, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.030886022374033928, + "learning_rate": 4.364105092145377e-05, + "loss": 2.7954399585723877, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.030112111940979958, + "learning_rate": 4.3511261076339e-05, + "loss": 2.78225040435791, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.03067166730761528, + "learning_rate": 4.3381680296614394e-05, + "loss": 2.8147826194763184, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.029821906238794327, + "learning_rate": 4.3252308686680626e-05, + "loss": 2.811027765274048, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.030433299019932747, + "learning_rate": 4.3123146350770146e-05, + "loss": 2.8178303241729736, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.02960469201207161, + "learning_rate": 4.2994193392946726e-05, + "loss": 2.802370548248291, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.029496584087610245, + "learning_rate": 4.286544991710541e-05, + "loss": 2.8156142234802246, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.029699044302105904, + "learning_rate": 4.2736916026972576e-05, + "loss": 2.7911806106567383, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.030225902795791626, + "learning_rate": 4.260859182610542e-05, + "loss": 2.7872819900512695, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.02939198724925518, + "learning_rate": 4.2480477417892776e-05, + "loss": 2.800001621246338, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.03059281036257744, + "learning_rate": 4.235257290555381e-05, + "loss": 2.8127222061157227, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.02927992306649685, + "learning_rate": 4.222487839213903e-05, + "loss": 2.8320517539978027, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.03038649819791317, + "learning_rate": 4.209739398052956e-05, + "loss": 2.8169116973876953, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.029806654900312424, + "learning_rate": 4.1970119773437316e-05, + "loss": 2.8279008865356445, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.029620900750160217, + "learning_rate": 4.184305587340483e-05, + "loss": 2.8212599754333496, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.029498178511857986, + "learning_rate": 4.171620238280511e-05, + "loss": 2.8112261295318604, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.02918926067650318, + "learning_rate": 4.158955940384179e-05, + "loss": 2.8290767669677734, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.029827842488884926, + "learning_rate": 4.146312703854881e-05, + "loss": 2.8396403789520264, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.029516329988837242, + "learning_rate": 4.133690538879046e-05, + "loss": 2.840583324432373, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.029057780280709267, + "learning_rate": 4.1210894556261226e-05, + "loss": 2.835113525390625, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.029345065355300903, + "learning_rate": 4.108509464248578e-05, + "loss": 2.8197970390319824, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.02873479388654232, + "learning_rate": 4.095950574881891e-05, + "loss": 2.8052587509155273, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.029737083241343498, + "learning_rate": 4.083412797644514e-05, + "loss": 2.819146156311035, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.0292612686753273, + "learning_rate": 4.0708961426379214e-05, + "loss": 2.8543701171875, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.028945421800017357, + "learning_rate": 4.0584006199465484e-05, + "loss": 2.804084300994873, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.029137294739484787, + "learning_rate": 4.0459262396378165e-05, + "loss": 2.7658708095550537, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.02911440096795559, + "learning_rate": 4.033473011762116e-05, + "loss": 2.8134429454803467, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.028932519257068634, + "learning_rate": 4.0210409463527656e-05, + "loss": 2.810455322265625, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.029404157772660255, + "learning_rate": 4.008630053426082e-05, + "loss": 2.802886962890625, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.028970446437597275, + "learning_rate": 3.996240342981279e-05, + "loss": 2.8426849842071533, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.029187830165028572, + "learning_rate": 3.983871825000529e-05, + "loss": 2.8366644382476807, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.029293285682797432, + "learning_rate": 3.971524509448925e-05, + "loss": 2.8261566162109375, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.029366638511419296, + "learning_rate": 3.959198406274472e-05, + "loss": 2.8125827312469482, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.02900587022304535, + "learning_rate": 3.946893525408097e-05, + "loss": 2.839684009552002, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.029162462800741196, + "learning_rate": 3.934609876763604e-05, + "loss": 2.8331997394561768, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.029060618951916695, + "learning_rate": 3.922347470237728e-05, + "loss": 2.8627192974090576, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.02887798845767975, + "learning_rate": 3.910106315710053e-05, + "loss": 2.859687089920044, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.02894580364227295, + "learning_rate": 3.897886423043061e-05, + "loss": 2.8195912837982178, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.029554955661296844, + "learning_rate": 3.8856878020820975e-05, + "loss": 2.852871894836426, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.02822914719581604, + "learning_rate": 3.873510462655375e-05, + "loss": 2.8671388626098633, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.02954651229083538, + "learning_rate": 3.861354414573954e-05, + "loss": 2.8524365425109863, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.02902315929532051, + "learning_rate": 3.849219667631735e-05, + "loss": 2.843954563140869, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.0290356632322073, + "learning_rate": 3.8371062316054764e-05, + "loss": 2.8504233360290527, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.029262971132993698, + "learning_rate": 3.8250141162547466e-05, + "loss": 2.8657913208007812, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.029469065368175507, + "learning_rate": 3.812943331321956e-05, + "loss": 2.82297945022583, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.029609832912683487, + "learning_rate": 3.8008938865322955e-05, + "loss": 2.856297016143799, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.029690174385905266, + "learning_rate": 3.788865791593813e-05, + "loss": 2.8302533626556396, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.03018389455974102, + "learning_rate": 3.7768590561973064e-05, + "loss": 2.8542540073394775, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.029053693637251854, + "learning_rate": 3.764873690016394e-05, + "loss": 2.8432576656341553, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.028549278154969215, + "learning_rate": 3.7529097027074654e-05, + "loss": 2.8324697017669678, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.030014613643288612, + "learning_rate": 3.7409671039096936e-05, + "loss": 2.800536870956421, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.02949996292591095, + "learning_rate": 3.7290459032450184e-05, + "loss": 2.835141897201538, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.028825247660279274, + "learning_rate": 3.717146110318123e-05, + "loss": 2.8263447284698486, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.029436098411679268, + "learning_rate": 3.7052677347164614e-05, + "loss": 2.8522257804870605, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.029510991647839546, + "learning_rate": 3.693410786010231e-05, + "loss": 2.851526975631714, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.028975380584597588, + "learning_rate": 3.6815752737523537e-05, + "loss": 2.820462465286255, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.029128702357411385, + "learning_rate": 3.6697612074784944e-05, + "loss": 2.861868381500244, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.02951551228761673, + "learning_rate": 3.657968596707029e-05, + "loss": 2.875210762023926, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.028196118772029877, + "learning_rate": 3.64619745093906e-05, + "loss": 2.8345048427581787, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.029215430840849876, + "learning_rate": 3.6344477796583714e-05, + "loss": 2.84328556060791, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.02857460267841816, + "learning_rate": 3.622719592331472e-05, + "loss": 2.8448305130004883, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.02842608466744423, + "learning_rate": 3.611012898407544e-05, + "loss": 2.860792636871338, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.03630338981747627, + "learning_rate": 3.599327707318464e-05, + "loss": 2.8723089694976807, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 4.522201061248779, + "eval_runtime": 57.3764, + "eval_samples_per_second": 42.561, + "eval_steps_per_second": 1.342, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.030005959793925285, + "learning_rate": 3.587664028478779e-05, + "loss": 2.7629683017730713, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.028917059302330017, + "learning_rate": 3.5760218712857046e-05, + "loss": 2.7894129753112793, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.029286719858646393, + "learning_rate": 3.564401245119119e-05, + "loss": 2.8119699954986572, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.02967856638133526, + "learning_rate": 3.552802159341543e-05, + "loss": 2.8115768432617188, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.029460743069648743, + "learning_rate": 3.541224623298155e-05, + "loss": 2.7809553146362305, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.028186975046992302, + "learning_rate": 3.529668646316765e-05, + "loss": 2.818805456161499, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.02975514344871044, + "learning_rate": 3.5181342377078166e-05, + "loss": 2.816215991973877, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.029532121494412422, + "learning_rate": 3.5066214067643785e-05, + "loss": 2.782461166381836, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.02879161946475506, + "learning_rate": 3.495130162762119e-05, + "loss": 2.768501043319702, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.029776159673929214, + "learning_rate": 3.483660514959343e-05, + "loss": 2.7995424270629883, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.028667304664850235, + "learning_rate": 3.4722124725969235e-05, + "loss": 2.773409843444824, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.028793245553970337, + "learning_rate": 3.460786044898346e-05, + "loss": 2.7938318252563477, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.029090600088238716, + "learning_rate": 3.4493812410696784e-05, + "loss": 2.823000907897949, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.028883792459964752, + "learning_rate": 3.437998070299564e-05, + "loss": 2.8175344467163086, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.028518877923488617, + "learning_rate": 3.426636541759225e-05, + "loss": 2.792815685272217, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.029769321903586388, + "learning_rate": 3.415296664602425e-05, + "loss": 2.831251382827759, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.029009172692894936, + "learning_rate": 3.403978447965507e-05, + "loss": 2.8345084190368652, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.028994156047701836, + "learning_rate": 3.392681900967355e-05, + "loss": 2.7925689220428467, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.029505716636776924, + "learning_rate": 3.381407032709395e-05, + "loss": 2.814758777618408, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.028683487325906754, + "learning_rate": 3.3701538522755714e-05, + "loss": 2.7954044342041016, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.029633600264787674, + "learning_rate": 3.35892236873238e-05, + "loss": 2.794569969177246, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.028943300247192383, + "learning_rate": 3.3477125911288284e-05, + "loss": 2.803800106048584, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.029026521369814873, + "learning_rate": 3.3365245284964225e-05, + "loss": 2.760383129119873, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.029057513922452927, + "learning_rate": 3.3253581898491836e-05, + "loss": 2.8081226348876953, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.02906658686697483, + "learning_rate": 3.3142135841836276e-05, + "loss": 2.8183770179748535, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.02926407754421234, + "learning_rate": 3.303090720478773e-05, + "loss": 2.831427574157715, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.029171453788876534, + "learning_rate": 3.291989607696092e-05, + "loss": 2.826587200164795, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.029656723141670227, + "learning_rate": 3.280910254779565e-05, + "loss": 2.7633590698242188, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.029490994289517403, + "learning_rate": 3.269852670655629e-05, + "loss": 2.7932424545288086, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.0291811041533947, + "learning_rate": 3.258816864233169e-05, + "loss": 2.8346633911132812, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.029164446517825127, + "learning_rate": 3.247802844403536e-05, + "loss": 2.8193140029907227, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.029394259676337242, + "learning_rate": 3.236810620040532e-05, + "loss": 2.8272342681884766, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.029221080243587494, + "learning_rate": 3.225840200000396e-05, + "loss": 2.79146146774292, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.028165455907583237, + "learning_rate": 3.214891593121788e-05, + "loss": 2.7826766967773438, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.0296048391610384, + "learning_rate": 3.203964808225801e-05, + "loss": 2.8119699954986572, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.028316112235188484, + "learning_rate": 3.193059854115968e-05, + "loss": 2.824860095977783, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02878737449645996, + "learning_rate": 3.182176739578195e-05, + "loss": 2.811997890472412, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029061628505587578, + "learning_rate": 3.171315473380815e-05, + "loss": 2.8370048999786377, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028607595711946487, + "learning_rate": 3.160476064274555e-05, + "loss": 2.8093791007995605, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.028144560754299164, + "learning_rate": 3.149658520992537e-05, + "loss": 2.8236052989959717, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.028813332319259644, + "learning_rate": 3.138862852250249e-05, + "loss": 2.838690996170044, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.028437303379178047, + "learning_rate": 3.1280890667455694e-05, + "loss": 2.800854206085205, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.028567243367433548, + "learning_rate": 3.1173371731587496e-05, + "loss": 2.8301069736480713, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.029187796637415886, + "learning_rate": 3.1066071801523905e-05, + "loss": 2.823208808898926, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.028203707188367844, + "learning_rate": 3.095899096371456e-05, + "loss": 2.804643154144287, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.02822236344218254, + "learning_rate": 3.0852129304432506e-05, + "loss": 2.8132784366607666, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.028921568766236305, + "learning_rate": 3.074548690977434e-05, + "loss": 2.8215417861938477, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.02867986261844635, + "learning_rate": 3.0639063865659834e-05, + "loss": 2.8318984508514404, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.028849760070443153, + "learning_rate": 3.0532860257832144e-05, + "loss": 2.8209228515625, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.028631621971726418, + "learning_rate": 3.04268761718576e-05, + "loss": 2.8219287395477295, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.028188595548272133, + "learning_rate": 3.0321111693125648e-05, + "loss": 2.814027786254883, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.028504883870482445, + "learning_rate": 3.0215566906848906e-05, + "loss": 2.847404956817627, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.02826615609228611, + "learning_rate": 3.0110241898062726e-05, + "loss": 2.808562755584717, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02844209223985672, + "learning_rate": 3.0005136751625748e-05, + "loss": 2.827244520187378, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.028765909373760223, + "learning_rate": 2.9900251552219178e-05, + "loss": 2.81117844581604, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.028130069375038147, + "learning_rate": 2.9795586384347202e-05, + "loss": 2.852351188659668, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.028408419340848923, + "learning_rate": 2.9691141332336494e-05, + "loss": 2.847090244293213, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.028523078188300133, + "learning_rate": 2.9586916480336735e-05, + "loss": 2.821716785430908, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.028461646288633347, + "learning_rate": 2.9482911912320006e-05, + "loss": 2.816019296646118, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.028531840071082115, + "learning_rate": 2.9379127712080756e-05, + "loss": 2.8226165771484375, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.028871113434433937, + "learning_rate": 2.9275563963236107e-05, + "loss": 2.831231117248535, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.0285959430038929, + "learning_rate": 2.9172220749225548e-05, + "loss": 2.8257393836975098, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02841980569064617, + "learning_rate": 2.906909815331085e-05, + "loss": 2.8298444747924805, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.028909973800182343, + "learning_rate": 2.896619625857588e-05, + "loss": 2.808438777923584, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.02884981967508793, + "learning_rate": 2.8863515147927003e-05, + "loss": 2.815157890319824, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.02836848609149456, + "learning_rate": 2.8761054904092514e-05, + "loss": 2.792750835418701, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.028464308008551598, + "learning_rate": 2.8658815609622665e-05, + "loss": 2.7594008445739746, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.02909877710044384, + "learning_rate": 2.855679734688993e-05, + "loss": 2.844888925552368, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.0285524632781744, + "learning_rate": 2.8455000198088502e-05, + "loss": 2.8162052631378174, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.02916364185512066, + "learning_rate": 2.8353424245234582e-05, + "loss": 2.8175275325775146, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.028360974043607712, + "learning_rate": 2.825206957016599e-05, + "loss": 2.8322806358337402, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.0366419292986393, + "learning_rate": 2.8150936254542376e-05, + "loss": 2.8780605792999268, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 4.539337635040283, + "eval_runtime": 57.482, + "eval_samples_per_second": 42.483, + "eval_steps_per_second": 1.34, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030007537454366684, + "learning_rate": 2.8050024379845118e-05, + "loss": 2.7774059772491455, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.028395716100931168, + "learning_rate": 2.7949334027376986e-05, + "loss": 2.8240408897399902, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.029259607195854187, + "learning_rate": 2.7848865278262427e-05, + "loss": 2.7780704498291016, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.029434163123369217, + "learning_rate": 2.7748618213447268e-05, + "loss": 2.802172899246216, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.02914051152765751, + "learning_rate": 2.764859291369885e-05, + "loss": 2.746731758117676, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.02971787191927433, + "learning_rate": 2.7548789459605603e-05, + "loss": 2.7998170852661133, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.029050175100564957, + "learning_rate": 2.744920793157743e-05, + "loss": 2.786863327026367, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.030421903356909752, + "learning_rate": 2.7349848409845462e-05, + "loss": 2.789545774459839, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.029742848128080368, + "learning_rate": 2.725071097446174e-05, + "loss": 2.827561855316162, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.03001682087779045, + "learning_rate": 2.715179570529959e-05, + "loss": 2.8226547241210938, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.030296841636300087, + "learning_rate": 2.7053102682053134e-05, + "loss": 2.795665740966797, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.02908152900636196, + "learning_rate": 2.6954631984237766e-05, + "loss": 2.7598729133605957, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.029059268534183502, + "learning_rate": 2.6856383691189347e-05, + "loss": 2.786234140396118, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.02878507599234581, + "learning_rate": 2.675835788206485e-05, + "loss": 2.7754065990448, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.02881552092730999, + "learning_rate": 2.6660554635841912e-05, + "loss": 2.779616355895996, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.02847588248550892, + "learning_rate": 2.6562974031318788e-05, + "loss": 2.7556540966033936, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.028446536511182785, + "learning_rate": 2.6465616147114485e-05, + "loss": 2.7858777046203613, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.028488771989941597, + "learning_rate": 2.6368481061668344e-05, + "loss": 2.8308897018432617, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.028518928214907646, + "learning_rate": 2.6271568853240557e-05, + "loss": 2.798438310623169, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.02807544358074665, + "learning_rate": 2.6174879599911347e-05, + "loss": 2.7854204177856445, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.028449799865484238, + "learning_rate": 2.6078413379581627e-05, + "loss": 2.787806510925293, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.02806367352604866, + "learning_rate": 2.5982170269972397e-05, + "loss": 2.773944139480591, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.028127942234277725, + "learning_rate": 2.588615034862507e-05, + "loss": 2.795468807220459, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.028206733986735344, + "learning_rate": 2.5790353692901158e-05, + "loss": 2.794552803039551, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.02769204042851925, + "learning_rate": 2.5694780379982238e-05, + "loss": 2.78456974029541, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.028235256671905518, + "learning_rate": 2.5599430486870055e-05, + "loss": 2.8258988857269287, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.028410933911800385, + "learning_rate": 2.5504304090386225e-05, + "loss": 2.8145394325256348, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.028073569759726524, + "learning_rate": 2.5409401267172474e-05, + "loss": 2.739804744720459, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.027985481545329094, + "learning_rate": 2.5314722093690176e-05, + "loss": 2.7385730743408203, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.0285568255931139, + "learning_rate": 2.522026664622071e-05, + "loss": 2.790156364440918, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.028381718322634697, + "learning_rate": 2.5126035000865124e-05, + "loss": 2.7715389728546143, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028455479070544243, + "learning_rate": 2.503202723354413e-05, + "loss": 2.804957628250122, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.028330320492386818, + "learning_rate": 2.4938243419998047e-05, + "loss": 2.813432216644287, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.02832239493727684, + "learning_rate": 2.484468363578689e-05, + "loss": 2.8087422847747803, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.028423218056559563, + "learning_rate": 2.475134795629004e-05, + "loss": 2.790351390838623, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.028114303946495056, + "learning_rate": 2.4658236456706357e-05, + "loss": 2.8367419242858887, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.028525900095701218, + "learning_rate": 2.456534921205403e-05, + "loss": 2.8024988174438477, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.027810173109173775, + "learning_rate": 2.4472686297170822e-05, + "loss": 2.8156440258026123, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.028172023594379425, + "learning_rate": 2.4380247786713403e-05, + "loss": 2.801828384399414, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.027999311685562134, + "learning_rate": 2.428803375515788e-05, + "loss": 2.7917897701263428, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.02863648533821106, + "learning_rate": 2.4196044276799397e-05, + "loss": 2.7998974323272705, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.027806704863905907, + "learning_rate": 2.410427942575228e-05, + "loss": 2.7955031394958496, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.028083806857466698, + "learning_rate": 2.4012739275949746e-05, + "loss": 2.8160462379455566, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02817711979150772, + "learning_rate": 2.3921423901144008e-05, + "loss": 2.7673726081848145, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.027995774522423744, + "learning_rate": 2.3830333374906342e-05, + "loss": 2.787623882293701, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.028293436393141747, + "learning_rate": 2.3739467770626637e-05, + "loss": 2.848489284515381, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.028710173442959785, + "learning_rate": 2.3648827161513686e-05, + "loss": 2.8186140060424805, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.028271062299609184, + "learning_rate": 2.3558411620595038e-05, + "loss": 2.822369337081909, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.02830388769507408, + "learning_rate": 2.3468221220716844e-05, + "loss": 2.8050756454467773, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.02831537462770939, + "learning_rate": 2.33782560345438e-05, + "loss": 2.8161096572875977, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.028237098827958107, + "learning_rate": 2.328851613455932e-05, + "loss": 2.7897520065307617, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.027944844216108322, + "learning_rate": 2.3199001593065204e-05, + "loss": 2.816229820251465, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.028575003147125244, + "learning_rate": 2.3109712482181684e-05, + "loss": 2.8069968223571777, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.028569547459483147, + "learning_rate": 2.3020648873847414e-05, + "loss": 2.78981614112854, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.028383128345012665, + "learning_rate": 2.2931810839819235e-05, + "loss": 2.820969343185425, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.027872422710061073, + "learning_rate": 2.284319845167253e-05, + "loss": 2.781301498413086, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.028342951089143753, + "learning_rate": 2.2754811780800532e-05, + "loss": 2.816469192504883, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.028373923152685165, + "learning_rate": 2.2666650898414827e-05, + "loss": 2.799290657043457, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.02807556465268135, + "learning_rate": 2.2578715875545095e-05, + "loss": 2.814924716949463, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.02789521962404251, + "learning_rate": 2.2491006783038976e-05, + "loss": 2.7950439453125, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.028297580778598785, + "learning_rate": 2.2403523691562092e-05, + "loss": 2.8535611629486084, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.02745051123201847, + "learning_rate": 2.2316266671598003e-05, + "loss": 2.7812037467956543, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.027246123179793358, + "learning_rate": 2.222923579344805e-05, + "loss": 2.778583288192749, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.027801746502518654, + "learning_rate": 2.2142431127231502e-05, + "loss": 2.812328815460205, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.027761677280068398, + "learning_rate": 2.2055852742885322e-05, + "loss": 2.829758882522583, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.028340548276901245, + "learning_rate": 2.196950071016409e-05, + "loss": 2.8203043937683105, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.027355236932635307, + "learning_rate": 2.1883375098640134e-05, + "loss": 2.777925968170166, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027668653056025505, + "learning_rate": 2.1797475977703352e-05, + "loss": 2.789498805999756, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02823868952691555, + "learning_rate": 2.1711803416560964e-05, + "loss": 2.7765135765075684, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.02812831476330757, + "learning_rate": 2.16263574842379e-05, + "loss": 2.774756669998169, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.027789659798145294, + "learning_rate": 2.15411382495764e-05, + "loss": 2.8317346572875977, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.035616710782051086, + "learning_rate": 2.1456145781236057e-05, + "loss": 2.8123791217803955, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 4.558599948883057, + "eval_runtime": 56.3211, + "eval_samples_per_second": 43.359, + "eval_steps_per_second": 1.367, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.027842728421092033, + "learning_rate": 2.1371380147693752e-05, + "loss": 2.766806125640869, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.027614781633019447, + "learning_rate": 2.1286841417243662e-05, + "loss": 2.7721548080444336, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.027729114517569542, + "learning_rate": 2.1202529657997137e-05, + "loss": 2.775094509124756, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.027813846245408058, + "learning_rate": 2.1118444937882538e-05, + "loss": 2.7677454948425293, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.027814052999019623, + "learning_rate": 2.1034587324645456e-05, + "loss": 2.764620304107666, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.027886489406228065, + "learning_rate": 2.095095688584846e-05, + "loss": 2.736621141433716, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.027611345052719116, + "learning_rate": 2.086755368887107e-05, + "loss": 2.775693893432617, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.027603253722190857, + "learning_rate": 2.0784377800909802e-05, + "loss": 2.7538721561431885, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.027774354442954063, + "learning_rate": 2.0701429288977847e-05, + "loss": 2.7577450275421143, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.028701378032565117, + "learning_rate": 2.0618708219905455e-05, + "loss": 2.792163848876953, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.02745729312300682, + "learning_rate": 2.0536214660339398e-05, + "loss": 2.7608468532562256, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.02753688395023346, + "learning_rate": 2.0453948676743267e-05, + "loss": 2.7727649211883545, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.027698034420609474, + "learning_rate": 2.037191033539731e-05, + "loss": 2.771090269088745, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.028323858976364136, + "learning_rate": 2.029009970239831e-05, + "loss": 2.7989578247070312, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.027763206511735916, + "learning_rate": 2.0208516843659678e-05, + "loss": 2.8134682178497314, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.027521144598722458, + "learning_rate": 2.012716182491114e-05, + "loss": 2.827221393585205, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.02790055237710476, + "learning_rate": 2.0046034711699048e-05, + "loss": 2.7750606536865234, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.027743415907025337, + "learning_rate": 1.9965135569386032e-05, + "loss": 2.754495620727539, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.02780855819582939, + "learning_rate": 1.9884464463151077e-05, + "loss": 2.793025255203247, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.027743151411414146, + "learning_rate": 1.980402145798933e-05, + "loss": 2.7627084255218506, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.02772294543683529, + "learning_rate": 1.9723806618712484e-05, + "loss": 2.810464382171631, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.027665726840496063, + "learning_rate": 1.9643820009947983e-05, + "loss": 2.7772934436798096, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.027962513267993927, + "learning_rate": 1.9564061696139687e-05, + "loss": 2.791640520095825, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.028066501021385193, + "learning_rate": 1.948453174154744e-05, + "loss": 2.7920327186584473, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.027441052719950676, + "learning_rate": 1.9405230210247042e-05, + "loss": 2.797529458999634, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.02827683836221695, + "learning_rate": 1.9326157166130365e-05, + "loss": 2.8248825073242188, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.027658650651574135, + "learning_rate": 1.9247312672905017e-05, + "loss": 2.8131494522094727, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.0280227642506361, + "learning_rate": 1.9168696794094758e-05, + "loss": 2.7728159427642822, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.02791561745107174, + "learning_rate": 1.9090309593038812e-05, + "loss": 2.7671456336975098, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.02787322923541069, + "learning_rate": 1.9012151132892425e-05, + "loss": 2.7740323543548584, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.027309276163578033, + "learning_rate": 1.893422147662644e-05, + "loss": 2.7773780822753906, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.027728214859962463, + "learning_rate": 1.8856520687027325e-05, + "loss": 2.7922744750976562, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.027990298345685005, + "learning_rate": 1.8779048826697318e-05, + "loss": 2.8120408058166504, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.02773648127913475, + "learning_rate": 1.8701805958054e-05, + "loss": 2.7819745540618896, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027117537334561348, + "learning_rate": 1.8624792143330544e-05, + "loss": 2.781357765197754, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.027264263480901718, + "learning_rate": 1.8548007444575656e-05, + "loss": 2.7825074195861816, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.027638141065835953, + "learning_rate": 1.8471451923653322e-05, + "loss": 2.809204578399658, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.027160588651895523, + "learning_rate": 1.8395125642242987e-05, + "loss": 2.784580707550049, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.02765049785375595, + "learning_rate": 1.8319028661839343e-05, + "loss": 2.7796647548675537, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027509981766343117, + "learning_rate": 1.8243161043752384e-05, + "loss": 2.802097797393799, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.02717861346900463, + "learning_rate": 1.8167522849107228e-05, + "loss": 2.790086269378662, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.027462702244520187, + "learning_rate": 1.809211413884418e-05, + "loss": 2.779426336288452, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.027430517598986626, + "learning_rate": 1.8016934973718762e-05, + "loss": 2.7842659950256348, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.027589300647377968, + "learning_rate": 1.7941985414301403e-05, + "loss": 2.812390089035034, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.02773849479854107, + "learning_rate": 1.786726552097766e-05, + "loss": 2.7788920402526855, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.027417341247200966, + "learning_rate": 1.7792775353947908e-05, + "loss": 2.766779899597168, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.027108531445264816, + "learning_rate": 1.7718514973227634e-05, + "loss": 2.774667501449585, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.02787691541016102, + "learning_rate": 1.7644484438647057e-05, + "loss": 2.8050103187561035, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.028047434985637665, + "learning_rate": 1.75706838098512e-05, + "loss": 2.795063018798828, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.027433287352323532, + "learning_rate": 1.749711314629992e-05, + "loss": 2.802816390991211, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.027421964332461357, + "learning_rate": 1.742377250726779e-05, + "loss": 2.8223254680633545, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.027546996250748634, + "learning_rate": 1.7350661951844065e-05, + "loss": 2.780585765838623, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.027431374415755272, + "learning_rate": 1.727778153893248e-05, + "loss": 2.787783145904541, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.027354050427675247, + "learning_rate": 1.720513132725162e-05, + "loss": 2.7926011085510254, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.02790139429271221, + "learning_rate": 1.713271137533433e-05, + "loss": 2.7482681274414062, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.027517175301909447, + "learning_rate": 1.7060521741528127e-05, + "loss": 2.791224956512451, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.027358947321772575, + "learning_rate": 1.698856248399482e-05, + "loss": 2.774211883544922, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.027352966368198395, + "learning_rate": 1.6916833660710785e-05, + "loss": 2.8025760650634766, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.02778322622179985, + "learning_rate": 1.6845335329466606e-05, + "loss": 2.7977163791656494, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.027194228023290634, + "learning_rate": 1.6774067547867113e-05, + "loss": 2.7744717597961426, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.027206066995859146, + "learning_rate": 1.6703030373331585e-05, + "loss": 2.7848451137542725, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.027613313868641853, + "learning_rate": 1.6632223863093328e-05, + "loss": 2.795206308364868, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.027662355452775955, + "learning_rate": 1.656164807419988e-05, + "loss": 2.771212577819824, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.027272798120975494, + "learning_rate": 1.6491303063512866e-05, + "loss": 2.850273609161377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.027627959847450256, + "learning_rate": 1.6421188887708013e-05, + "loss": 2.7897393703460693, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.027814874425530434, + "learning_rate": 1.6351305603275067e-05, + "loss": 2.8047971725463867, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.027428552508354187, + "learning_rate": 1.6281653266517658e-05, + "loss": 2.7656729221343994, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.02777225710451603, + "learning_rate": 1.6212231933553485e-05, + "loss": 2.7727203369140625, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.02802879735827446, + "learning_rate": 1.6143041660313997e-05, + "loss": 2.847011089324951, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.02760482393205166, + "learning_rate": 1.6074082502544668e-05, + "loss": 2.7733654975891113, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.02776470221579075, + "learning_rate": 1.6005354515804503e-05, + "loss": 2.7941837310791016, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.03535662591457367, + "learning_rate": 1.59368577554664e-05, + "loss": 2.787112236022949, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 4.572118282318115, + "eval_runtime": 56.9428, + "eval_samples_per_second": 42.885, + "eval_steps_per_second": 1.352, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.026943478733301163, + "learning_rate": 1.5868592276717137e-05, + "loss": 2.7454490661621094, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.027496816590428352, + "learning_rate": 1.5800558134556806e-05, + "loss": 2.783641815185547, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.026721451431512833, + "learning_rate": 1.5732755383799315e-05, + "loss": 2.7537829875946045, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.027940884232521057, + "learning_rate": 1.5665184079072162e-05, + "loss": 2.7764663696289062, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.027781952172517776, + "learning_rate": 1.5597844274816352e-05, + "loss": 2.776278495788574, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.027947992086410522, + "learning_rate": 1.553073602528626e-05, + "loss": 2.7712326049804688, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.027257733047008514, + "learning_rate": 1.546385938454982e-05, + "loss": 2.774122714996338, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.026782797649502754, + "learning_rate": 1.5397214406488374e-05, + "loss": 2.7600014209747314, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.02728961408138275, + "learning_rate": 1.533080114479657e-05, + "loss": 2.7231063842773438, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.027733273804187775, + "learning_rate": 1.5264619652982352e-05, + "loss": 2.807209014892578, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.02731870673596859, + "learning_rate": 1.5198669984366876e-05, + "loss": 2.7994370460510254, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.02707100473344326, + "learning_rate": 1.5132952192084793e-05, + "loss": 2.762448787689209, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.027301348745822906, + "learning_rate": 1.5067466329083533e-05, + "loss": 2.754739284515381, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.02716146782040596, + "learning_rate": 1.5002212448123997e-05, + "loss": 2.7493300437927246, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.02700486034154892, + "learning_rate": 1.4937190601780052e-05, + "loss": 2.767707347869873, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.0270010344684124, + "learning_rate": 1.4872400842438576e-05, + "loss": 2.7867441177368164, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.02769402787089348, + "learning_rate": 1.4807843222299557e-05, + "loss": 2.7755560874938965, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.027279293164610863, + "learning_rate": 1.4743517793375814e-05, + "loss": 2.7624804973602295, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0273398756980896, + "learning_rate": 1.4679424607493315e-05, + "loss": 2.7540690898895264, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.026957856491208076, + "learning_rate": 1.4615563716290698e-05, + "loss": 2.774245500564575, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.028008051216602325, + "learning_rate": 1.455193517121951e-05, + "loss": 2.768254280090332, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.027576304972171783, + "learning_rate": 1.448853902354421e-05, + "loss": 2.7471680641174316, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.02723308838903904, + "learning_rate": 1.4425375324341827e-05, + "loss": 2.782475233078003, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.028015952557325363, + "learning_rate": 1.4362444124502309e-05, + "loss": 2.761064052581787, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.02686639130115509, + "learning_rate": 1.429974547472814e-05, + "loss": 2.780773162841797, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.027318131178617477, + "learning_rate": 1.4237279425534442e-05, + "loss": 2.790548801422119, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.027507461607456207, + "learning_rate": 1.4175046027249065e-05, + "loss": 2.7799487113952637, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.027209127321839333, + "learning_rate": 1.4113045330012324e-05, + "loss": 2.7879908084869385, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.027000298723578453, + "learning_rate": 1.4051277383776953e-05, + "loss": 2.785269260406494, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.026835331693291664, + "learning_rate": 1.398974223830839e-05, + "loss": 2.7634077072143555, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.027131548151373863, + "learning_rate": 1.3928439943184394e-05, + "loss": 2.780407667160034, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.027154795825481415, + "learning_rate": 1.3867370547795065e-05, + "loss": 2.7430217266082764, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.027125395834445953, + "learning_rate": 1.3806534101342891e-05, + "loss": 2.7663848400115967, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.02727346308529377, + "learning_rate": 1.3745930652842728e-05, + "loss": 2.7848246097564697, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.027327269315719604, + "learning_rate": 1.368556025112176e-05, + "loss": 2.7878546714782715, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.027070313692092896, + "learning_rate": 1.362542294481918e-05, + "loss": 2.768770933151245, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.026788033545017242, + "learning_rate": 1.356551878238656e-05, + "loss": 2.776149034500122, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.027060916647315025, + "learning_rate": 1.3505847812087724e-05, + "loss": 2.7886319160461426, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.027068400755524635, + "learning_rate": 1.3446410081998384e-05, + "loss": 2.750335216522217, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.02673197351396084, + "learning_rate": 1.3387205640006494e-05, + "loss": 2.796083450317383, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.026712248101830482, + "learning_rate": 1.332823453381196e-05, + "loss": 2.7682032585144043, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.02712504006922245, + "learning_rate": 1.3269496810926807e-05, + "loss": 2.790653705596924, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.027259107679128647, + "learning_rate": 1.3210992518674848e-05, + "loss": 2.7920985221862793, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.02679583430290222, + "learning_rate": 1.315272170419201e-05, + "loss": 2.7540664672851562, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.026917198672890663, + "learning_rate": 1.3094684414426073e-05, + "loss": 2.7794861793518066, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.02691952884197235, + "learning_rate": 1.3036880696136532e-05, + "loss": 2.7636754512786865, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.027165602892637253, + "learning_rate": 1.2979310595894882e-05, + "loss": 2.8022518157958984, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.0270492322742939, + "learning_rate": 1.2921974160084251e-05, + "loss": 2.763423204421997, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.0269774217158556, + "learning_rate": 1.2864871434899672e-05, + "loss": 2.7642290592193604, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.02735322155058384, + "learning_rate": 1.2808002466347654e-05, + "loss": 2.7912001609802246, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.026876460760831833, + "learning_rate": 1.2751367300246535e-05, + "loss": 2.8102259635925293, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.027073487639427185, + "learning_rate": 1.2694965982226301e-05, + "loss": 2.7805755138397217, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.026842350140213966, + "learning_rate": 1.2638798557728402e-05, + "loss": 2.80954909324646, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.027058839797973633, + "learning_rate": 1.2582865072006e-05, + "loss": 2.79652738571167, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.026913955807685852, + "learning_rate": 1.2527165570123586e-05, + "loss": 2.7777507305145264, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026756104081869125, + "learning_rate": 1.2471700096957376e-05, + "loss": 2.7783961296081543, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.026751704514026642, + "learning_rate": 1.2416468697194767e-05, + "loss": 2.765263080596924, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.02720797434449196, + "learning_rate": 1.2361471415334795e-05, + "loss": 2.7831039428710938, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.027288692072033882, + "learning_rate": 1.2306708295687706e-05, + "loss": 2.780830144882202, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.027102869004011154, + "learning_rate": 1.2252179382375211e-05, + "loss": 2.7735419273376465, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.027081217616796494, + "learning_rate": 1.2197884719330281e-05, + "loss": 2.761964797973633, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.02710801735520363, + "learning_rate": 1.2143824350297104e-05, + "loss": 2.7670583724975586, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.027270127087831497, + "learning_rate": 1.2089998318831137e-05, + "loss": 2.786961078643799, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.02709820494055748, + "learning_rate": 1.2036406668299062e-05, + "loss": 2.796149969100952, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.027192605659365654, + "learning_rate": 1.198304944187868e-05, + "loss": 2.7829535007476807, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.026996217668056488, + "learning_rate": 1.1929926682558995e-05, + "loss": 2.801142930984497, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.027222832664847374, + "learning_rate": 1.1877038433139998e-05, + "loss": 2.773264169692993, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.027082091197371483, + "learning_rate": 1.1824384736232879e-05, + "loss": 2.8198862075805664, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.027131790295243263, + "learning_rate": 1.1771965634259653e-05, + "loss": 2.8299996852874756, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.027189496904611588, + "learning_rate": 1.1719781169453536e-05, + "loss": 2.7663733959198, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.02728065848350525, + "learning_rate": 1.1667831383858562e-05, + "loss": 2.8235511779785156, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.035003211349248886, + "learning_rate": 1.1616116319329808e-05, + "loss": 2.812906265258789, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 4.58276891708374, + "eval_runtime": 57.7425, + "eval_samples_per_second": 42.291, + "eval_steps_per_second": 1.334, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.026835810393095016, + "learning_rate": 1.1564636017533036e-05, + "loss": 2.7499475479125977, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.02696690894663334, + "learning_rate": 1.151339051994508e-05, + "loss": 2.7551589012145996, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.026555679738521576, + "learning_rate": 1.146237986785356e-05, + "loss": 2.753525733947754, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.02668551355600357, + "learning_rate": 1.141160410235674e-05, + "loss": 2.7452681064605713, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.027071513235569, + "learning_rate": 1.1361063264363776e-05, + "loss": 2.77901029586792, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.027035463601350784, + "learning_rate": 1.1310757394594514e-05, + "loss": 2.769394874572754, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.026944896206259727, + "learning_rate": 1.1260686533579481e-05, + "loss": 2.78621768951416, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.026308422908186913, + "learning_rate": 1.121085072165986e-05, + "loss": 2.746005058288574, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.02692459337413311, + "learning_rate": 1.1161249998987429e-05, + "loss": 2.7592382431030273, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.026701323688030243, + "learning_rate": 1.111188440552471e-05, + "loss": 2.77679181098938, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.02708185277879238, + "learning_rate": 1.106275398104455e-05, + "loss": 2.782973289489746, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.026750970631837845, + "learning_rate": 1.1013858765130468e-05, + "loss": 2.760465621948242, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.026963144540786743, + "learning_rate": 1.0965198797176513e-05, + "loss": 2.7820069789886475, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.02676742523908615, + "learning_rate": 1.0916774116387103e-05, + "loss": 2.7657270431518555, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.027025895193219185, + "learning_rate": 1.0868584761777125e-05, + "loss": 2.7613558769226074, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.027024701237678528, + "learning_rate": 1.0820630772171869e-05, + "loss": 2.7726471424102783, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.027291523292660713, + "learning_rate": 1.0772912186206987e-05, + "loss": 2.807295799255371, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.02704339101910591, + "learning_rate": 1.0725429042328542e-05, + "loss": 2.788114309310913, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.02788626216351986, + "learning_rate": 1.0678181378792794e-05, + "loss": 2.7785983085632324, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.02694671042263508, + "learning_rate": 1.0631169233666306e-05, + "loss": 2.75980281829834, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.02722201682627201, + "learning_rate": 1.0584392644826016e-05, + "loss": 2.7798352241516113, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.026429232209920883, + "learning_rate": 1.053785164995886e-05, + "loss": 2.746462821960449, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.027395036071538925, + "learning_rate": 1.0491546286562148e-05, + "loss": 2.7769758701324463, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.027433833107352257, + "learning_rate": 1.0445476591943259e-05, + "loss": 2.7528226375579834, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.026795532554388046, + "learning_rate": 1.0399642603219712e-05, + "loss": 2.7847824096679688, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.026764724403619766, + "learning_rate": 1.0354044357319129e-05, + "loss": 2.7527782917022705, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.026894114911556244, + "learning_rate": 1.0308681890979173e-05, + "loss": 2.7632064819335938, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.026841631159186363, + "learning_rate": 1.026355524074758e-05, + "loss": 2.755859851837158, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.02638275921344757, + "learning_rate": 1.0218664442982084e-05, + "loss": 2.7602035999298096, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.026950951665639877, + "learning_rate": 1.0174009533850368e-05, + "loss": 2.8016932010650635, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.02662278153002262, + "learning_rate": 1.0129590549330119e-05, + "loss": 2.7938461303710938, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.02673911489546299, + "learning_rate": 1.0085407525208865e-05, + "loss": 2.7740683555603027, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.02670123428106308, + "learning_rate": 1.0041460497084135e-05, + "loss": 2.771376132965088, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.026872053742408752, + "learning_rate": 9.997749500363172e-06, + "loss": 2.7445852756500244, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.026589123532176018, + "learning_rate": 9.954274570263191e-06, + "loss": 2.7571749687194824, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.02649911306798458, + "learning_rate": 9.911035741811153e-06, + "loss": 2.768310546875, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.0264277383685112, + "learning_rate": 9.868033049843787e-06, + "loss": 2.7388150691986084, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.02685769833624363, + "learning_rate": 9.825266529007601e-06, + "loss": 2.793008327484131, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.026673899963498116, + "learning_rate": 9.782736213758782e-06, + "loss": 2.8166663646698, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.026844056323170662, + "learning_rate": 9.740442138363317e-06, + "loss": 2.773449182510376, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.026580123230814934, + "learning_rate": 9.698384336896679e-06, + "loss": 2.7908055782318115, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.02664843201637268, + "learning_rate": 9.656562843244114e-06, + "loss": 2.7975826263427734, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.026540955528616905, + "learning_rate": 9.614977691100445e-06, + "loss": 2.7841620445251465, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.026925355195999146, + "learning_rate": 9.57362891397007e-06, + "loss": 2.7862730026245117, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.02668730728328228, + "learning_rate": 9.532516545167004e-06, + "loss": 2.7859320640563965, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.02650708518922329, + "learning_rate": 9.491640617814634e-06, + "loss": 2.793994426727295, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.0267381239682436, + "learning_rate": 9.451001164846092e-06, + "loss": 2.786421298980713, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.02687137760221958, + "learning_rate": 9.410598219003725e-06, + "loss": 2.752488851547241, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.02661444991827011, + "learning_rate": 9.370431812839556e-06, + "loss": 2.7778420448303223, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.026941899210214615, + "learning_rate": 9.330501978714857e-06, + "loss": 2.777524948120117, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.0268207136541605, + "learning_rate": 9.290808748800408e-06, + "loss": 2.7725157737731934, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.02680084854364395, + "learning_rate": 9.25135215507634e-06, + "loss": 2.7609915733337402, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.026696713641285896, + "learning_rate": 9.212132229332094e-06, + "loss": 2.7365245819091797, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.0265318863093853, + "learning_rate": 9.173149003166458e-06, + "loss": 2.7494451999664307, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.026489339768886566, + "learning_rate": 9.134402507987533e-06, + "loss": 2.7605128288269043, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.026866141706705093, + "learning_rate": 9.095892775012672e-06, + "loss": 2.7785377502441406, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.02668614685535431, + "learning_rate": 9.057619835268437e-06, + "loss": 2.7544281482696533, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.026771903038024902, + "learning_rate": 9.019583719590641e-06, + "loss": 2.7701380252838135, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.026549167931079865, + "learning_rate": 8.981784458624375e-06, + "loss": 2.7562901973724365, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.02666434273123741, + "learning_rate": 8.944222082823717e-06, + "loss": 2.786470890045166, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.026594869792461395, + "learning_rate": 8.906896622452055e-06, + "loss": 2.791085958480835, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.026912730187177658, + "learning_rate": 8.86980810758183e-06, + "loss": 2.7630703449249268, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.02666492760181427, + "learning_rate": 8.832956568094596e-06, + "loss": 2.771186351776123, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.02624955028295517, + "learning_rate": 8.796342033680928e-06, + "loss": 2.740208625793457, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.02678191289305687, + "learning_rate": 8.759964533840574e-06, + "loss": 2.7912135124206543, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.026293903589248657, + "learning_rate": 8.723824097882145e-06, + "loss": 2.7572717666625977, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.02646786905825138, + "learning_rate": 8.687920754923393e-06, + "loss": 2.7562999725341797, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.026905927807092667, + "learning_rate": 8.652254533890956e-06, + "loss": 2.7835705280303955, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.026924896985292435, + "learning_rate": 8.616825463520493e-06, + "loss": 2.7965805530548096, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.026362398639321327, + "learning_rate": 8.581633572356576e-06, + "loss": 2.737618923187256, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.026691535487771034, + "learning_rate": 8.546678888752629e-06, + "loss": 2.776054620742798, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.03468878194689751, + "learning_rate": 8.51196144087103e-06, + "loss": 2.7775015830993652, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 4.590993881225586, + "eval_runtime": 57.7996, + "eval_samples_per_second": 42.249, + "eval_steps_per_second": 1.332, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.026764800772070885, + "learning_rate": 8.477481256683005e-06, + "loss": 2.7140045166015625, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.026526780799031258, + "learning_rate": 8.44323836396864e-06, + "loss": 2.7631232738494873, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.026312852278351784, + "learning_rate": 8.409232790316763e-06, + "loss": 2.7715871334075928, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.026490503922104836, + "learning_rate": 8.375464563125094e-06, + "loss": 2.7789485454559326, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.02680407650768757, + "learning_rate": 8.341933709600102e-06, + "loss": 2.7781310081481934, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.026628121733665466, + "learning_rate": 8.308640256756916e-06, + "loss": 2.754638195037842, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.026735341176390648, + "learning_rate": 8.275584231419512e-06, + "loss": 2.7532246112823486, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.026857327669858932, + "learning_rate": 8.24276566022055e-06, + "loss": 2.753671646118164, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.026332298293709755, + "learning_rate": 8.21018456960132e-06, + "loss": 2.752696990966797, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.026672109961509705, + "learning_rate": 8.177840985811858e-06, + "loss": 2.8153340816497803, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.02677585557103157, + "learning_rate": 8.145734934910757e-06, + "loss": 2.7665958404541016, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.02687709592282772, + "learning_rate": 8.113866442765364e-06, + "loss": 2.7748546600341797, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.026170765981078148, + "learning_rate": 8.082235535051449e-06, + "loss": 2.759859561920166, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.026528827846050262, + "learning_rate": 8.050842237253533e-06, + "loss": 2.804813861846924, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.0264429971575737, + "learning_rate": 8.019686574664564e-06, + "loss": 2.765291213989258, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.026966489851474762, + "learning_rate": 7.988768572386171e-06, + "loss": 2.7834548950195312, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.02647687867283821, + "learning_rate": 7.958088255328415e-06, + "loss": 2.7787539958953857, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.026527805253863335, + "learning_rate": 7.927645648209837e-06, + "loss": 2.7454590797424316, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.026494361460208893, + "learning_rate": 7.897440775557506e-06, + "loss": 2.7637276649475098, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.0264204703271389, + "learning_rate": 7.867473661706946e-06, + "loss": 2.7916486263275146, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.02635893039405346, + "learning_rate": 7.837744330802136e-06, + "loss": 2.7680537700653076, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.026351723819971085, + "learning_rate": 7.808252806795446e-06, + "loss": 2.7503089904785156, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.026458928361535072, + "learning_rate": 7.778999113447673e-06, + "loss": 2.7462029457092285, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.026599271222949028, + "learning_rate": 7.749983274327994e-06, + "loss": 2.7531824111938477, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.02652306668460369, + "learning_rate": 7.721205312813921e-06, + "loss": 2.7552847862243652, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.026365090161561966, + "learning_rate": 7.69266525209132e-06, + "loss": 2.739084243774414, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.02636418491601944, + "learning_rate": 7.664363115154476e-06, + "loss": 2.72769832611084, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.026417085900902748, + "learning_rate": 7.636298924805872e-06, + "loss": 2.726992607116699, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.02646288275718689, + "learning_rate": 7.608472703656282e-06, + "loss": 2.7834439277648926, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.026408160105347633, + "learning_rate": 7.580884474124833e-06, + "loss": 2.773449659347534, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.026651300489902496, + "learning_rate": 7.55353425843888e-06, + "loss": 2.762749671936035, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.02692393772304058, + "learning_rate": 7.526422078633939e-06, + "loss": 2.7532958984375, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.026534877717494965, + "learning_rate": 7.4995479565538476e-06, + "loss": 2.7620131969451904, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.02664080075919628, + "learning_rate": 7.4729119138506044e-06, + "loss": 2.7795252799987793, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.02610137313604355, + "learning_rate": 7.446513971984366e-06, + "loss": 2.740626096725464, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.026373442262411118, + "learning_rate": 7.420354152223485e-06, + "loss": 2.809202194213867, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.026477070525288582, + "learning_rate": 7.394432475644402e-06, + "loss": 2.7980127334594727, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.026563815772533417, + "learning_rate": 7.368748963131854e-06, + "loss": 2.7525994777679443, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.026580089703202248, + "learning_rate": 7.343303635378537e-06, + "loss": 2.7786154747009277, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.026262899860739708, + "learning_rate": 7.31809651288524e-06, + "loss": 2.755073070526123, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.02683323062956333, + "learning_rate": 7.2931276159609475e-06, + "loss": 2.784115791320801, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.0266089029610157, + "learning_rate": 7.268396964722602e-06, + "loss": 2.757763385772705, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.0266721174120903, + "learning_rate": 7.243904579095308e-06, + "loss": 2.7321813106536865, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.026465903967618942, + "learning_rate": 7.219650478812065e-06, + "loss": 2.7546777725219727, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.02655668370425701, + "learning_rate": 7.195634683413996e-06, + "loss": 2.7383837699890137, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.02678738348186016, + "learning_rate": 7.171857212250225e-06, + "loss": 2.7965049743652344, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.02643296681344509, + "learning_rate": 7.148318084477766e-06, + "loss": 2.757577657699585, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.026728913187980652, + "learning_rate": 7.125017319061728e-06, + "loss": 2.7963476181030273, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.026841776445508003, + "learning_rate": 7.101954934775084e-06, + "loss": 2.761970043182373, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026342229917645454, + "learning_rate": 7.079130950198802e-06, + "loss": 2.7694449424743652, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.026661433279514313, + "learning_rate": 7.05654538372171e-06, + "loss": 2.761528491973877, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.026377758011221886, + "learning_rate": 7.034198253540665e-06, + "loss": 2.7743632793426514, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.02674027532339096, + "learning_rate": 7.012089577660255e-06, + "loss": 2.809512138366699, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.026375900954008102, + "learning_rate": 6.990219373893128e-06, + "loss": 2.7861762046813965, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.0262554120272398, + "learning_rate": 6.968587659859665e-06, + "loss": 2.7638261318206787, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.026556264609098434, + "learning_rate": 6.9471944529881376e-06, + "loss": 2.744332790374756, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.027108121663331985, + "learning_rate": 6.926039770514686e-06, + "loss": 2.7774176597595215, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.026768099516630173, + "learning_rate": 6.905123629483277e-06, + "loss": 2.7910962104797363, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.02630891092121601, + "learning_rate": 6.884446046745608e-06, + "loss": 2.7563364505767822, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.026759762316942215, + "learning_rate": 6.8640070389613055e-06, + "loss": 2.7848401069641113, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.02670259028673172, + "learning_rate": 6.843806622597661e-06, + "loss": 2.7814881801605225, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.026632139459252357, + "learning_rate": 6.823844813929796e-06, + "loss": 2.759962797164917, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.026562802493572235, + "learning_rate": 6.804121629040597e-06, + "loss": 2.7739615440368652, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.026842210441827774, + "learning_rate": 6.7846370838206455e-06, + "loss": 2.740704298019409, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.02690100111067295, + "learning_rate": 6.765391193968287e-06, + "loss": 2.790417194366455, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.026587538421154022, + "learning_rate": 6.7463839749896335e-06, + "loss": 2.7733266353607178, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.026449866592884064, + "learning_rate": 6.727615442198394e-06, + "loss": 2.759732484817505, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.026533210650086403, + "learning_rate": 6.709085610716107e-06, + "loss": 2.749570608139038, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.02660883031785488, + "learning_rate": 6.690794495471878e-06, + "loss": 2.7633883953094482, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.02642008289694786, + "learning_rate": 6.672742111202543e-06, + "loss": 2.7738819122314453, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.026371262967586517, + "learning_rate": 6.654928472452569e-06, + "loss": 2.750501871109009, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.034971874207258224, + "learning_rate": 6.637353593574088e-06, + "loss": 2.773648977279663, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 4.597776889801025, + "eval_runtime": 57.7413, + "eval_samples_per_second": 42.292, + "eval_steps_per_second": 1.334, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.026444269344210625, + "learning_rate": 6.620017488726867e-06, + "loss": 2.785430431365967, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.026068098843097687, + "learning_rate": 6.602920171878265e-06, + "loss": 2.7389187812805176, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.02634287439286709, + "learning_rate": 6.5860616568033465e-06, + "loss": 2.7681965827941895, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.02599756047129631, + "learning_rate": 6.569441957084671e-06, + "loss": 2.7339017391204834, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.026426687836647034, + "learning_rate": 6.5530610861124e-06, + "loss": 2.7654523849487305, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.026443755254149437, + "learning_rate": 6.536919057084357e-06, + "loss": 2.7667272090911865, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.02631179802119732, + "learning_rate": 6.521015883005866e-06, + "loss": 2.751314878463745, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.026355480775237083, + "learning_rate": 6.5053515766898195e-06, + "loss": 2.773685932159424, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.026273800060153008, + "learning_rate": 6.489926150756674e-06, + "loss": 2.7847769260406494, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.026335887610912323, + "learning_rate": 6.4747396176344176e-06, + "loss": 2.766859531402588, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.026322951540350914, + "learning_rate": 6.459791989558542e-06, + "loss": 2.780479907989502, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.02634085901081562, + "learning_rate": 6.445083278572134e-06, + "loss": 2.7381904125213623, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.026212027296423912, + "learning_rate": 6.430613496525685e-06, + "loss": 2.7327303886413574, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.026233792304992676, + "learning_rate": 6.416382655077248e-06, + "loss": 2.7877025604248047, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.026520710438489914, + "learning_rate": 6.402390765692381e-06, + "loss": 2.7866268157958984, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.026276713237166405, + "learning_rate": 6.388637839644073e-06, + "loss": 2.76708722114563, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.026201969012618065, + "learning_rate": 6.375123888012812e-06, + "loss": 2.7421412467956543, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.026183459907770157, + "learning_rate": 6.361848921686523e-06, + "loss": 2.766507148742676, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.0260463934391737, + "learning_rate": 6.348812951360662e-06, + "loss": 2.761979103088379, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.02621057629585266, + "learning_rate": 6.33601598753802e-06, + "loss": 2.739781141281128, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.026393795385956764, + "learning_rate": 6.323458040528923e-06, + "loss": 2.7508270740509033, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.026347093284130096, + "learning_rate": 6.311139120451062e-06, + "loss": 2.743001937866211, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.02633446641266346, + "learning_rate": 6.299059237229567e-06, + "loss": 2.7588772773742676, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.026023060083389282, + "learning_rate": 6.287218400596961e-06, + "loss": 2.7275609970092773, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.026417123153805733, + "learning_rate": 6.27561662009321e-06, + "loss": 2.757868766784668, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.026473915204405785, + "learning_rate": 6.264253905065675e-06, + "loss": 2.7649319171905518, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.02639617957174778, + "learning_rate": 6.2531302646690866e-06, + "loss": 2.740844488143921, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.026274150237441063, + "learning_rate": 6.242245707865545e-06, + "loss": 2.771275520324707, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.026055404916405678, + "learning_rate": 6.23160024342455e-06, + "loss": 2.791905403137207, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.025836700573563576, + "learning_rate": 6.221193879922969e-06, + "loss": 2.7217962741851807, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.026319170370697975, + "learning_rate": 6.211026625745005e-06, + "loss": 2.7746083736419678, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.026230650022625923, + "learning_rate": 6.201098489082232e-06, + "loss": 2.770860433578491, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.026214411482214928, + "learning_rate": 6.191409477933554e-06, + "loss": 2.7712812423706055, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.02619052492082119, + "learning_rate": 6.18195960010528e-06, + "loss": 2.7662360668182373, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.026257462799549103, + "learning_rate": 6.172748863210954e-06, + "loss": 2.7349135875701904, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.026416735723614693, + "learning_rate": 6.163777274671552e-06, + "loss": 2.757458448410034, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.026388030499219894, + "learning_rate": 6.155044841715289e-06, + "loss": 2.744500160217285, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.026406629011034966, + "learning_rate": 6.146551571377715e-06, + "loss": 2.7591867446899414, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.026343751698732376, + "learning_rate": 6.1382974705017456e-06, + "loss": 2.759531021118164, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.02655695006251335, + "learning_rate": 6.130282545737535e-06, + "loss": 2.7539916038513184, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.026398885995149612, + "learning_rate": 6.1225068035425725e-06, + "loss": 2.743812084197998, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.026152826845645905, + "learning_rate": 6.114970250181616e-06, + "loss": 2.7331364154815674, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.026342187076807022, + "learning_rate": 6.107672891726724e-06, + "loss": 2.7509233951568604, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.02646729163825512, + "learning_rate": 6.1006147340572916e-06, + "loss": 2.797689437866211, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.02662649005651474, + "learning_rate": 6.093795782859917e-06, + "loss": 2.757904529571533, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.026698576286435127, + "learning_rate": 6.087216043628502e-06, + "loss": 2.7585411071777344, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.026497380807995796, + "learning_rate": 6.0808755216642146e-06, + "loss": 2.759650945663452, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.026584723964333534, + "learning_rate": 6.07477422207556e-06, + "loss": 2.766024351119995, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.026535525918006897, + "learning_rate": 6.06891214977818e-06, + "loss": 2.7463855743408203, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.026703383773565292, + "learning_rate": 6.063289309495051e-06, + "loss": 2.7799477577209473, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.026495758444070816, + "learning_rate": 6.057905705756421e-06, + "loss": 2.767401695251465, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.02615673840045929, + "learning_rate": 6.052761342899737e-06, + "loss": 2.7813711166381836, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.02638433314859867, + "learning_rate": 6.0478562250697185e-06, + "loss": 2.788975477218628, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.026382725685834885, + "learning_rate": 6.043190356218355e-06, + "loss": 2.7315516471862793, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.02635836973786354, + "learning_rate": 6.038763740104835e-06, + "loss": 2.756807327270508, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.02632402814924717, + "learning_rate": 6.034576380295586e-06, + "loss": 2.754215717315674, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.026373671367764473, + "learning_rate": 6.030628280164338e-06, + "loss": 2.7971506118774414, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.026456575840711594, + "learning_rate": 6.026919442891958e-06, + "loss": 2.7842516899108887, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.026110582053661346, + "learning_rate": 6.023449871466615e-06, + "loss": 2.7473106384277344, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.02640441618859768, + "learning_rate": 6.020219568683678e-06, + "loss": 2.766808271408081, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.0265651848167181, + "learning_rate": 6.0172285371456925e-06, + "loss": 2.804321527481079, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.026652604341506958, + "learning_rate": 6.014476779262568e-06, + "loss": 2.775193214416504, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.026517583057284355, + "learning_rate": 6.011964297251285e-06, + "loss": 2.784888744354248, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02647361345589161, + "learning_rate": 6.0096910931361275e-06, + "loss": 2.7551116943359375, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.026397163048386574, + "learning_rate": 6.007657168748551e-06, + "loss": 2.747201442718506, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.02646862529218197, + "learning_rate": 6.005862525727277e-06, + "loss": 2.780600070953369, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.026226090267300606, + "learning_rate": 6.0043071655181986e-06, + "loss": 2.779751777648926, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.026499273255467415, + "learning_rate": 6.002991089374446e-06, + "loss": 2.7725064754486084, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.026265744119882584, + "learning_rate": 6.0019142983563875e-06, + "loss": 2.7508344650268555, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.026274647563695908, + "learning_rate": 6.001076793331555e-06, + "loss": 2.7637529373168945, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.026484506204724312, + "learning_rate": 6.000478574974689e-06, + "loss": 2.753441095352173, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.034853070974349976, + "learning_rate": 6.000119643767767e-06, + "loss": 2.804600477218628, + "step": 3600 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.534806680666112e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-3600/training_args.bin b/runs/l2r90-baseline/checkpoint-3600/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-3600/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-72/chat_template.jinja b/runs/l2r90-baseline/checkpoint-72/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-72/config.json b/runs/l2r90-baseline/checkpoint-72/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-72/generation_config.json b/runs/l2r90-baseline/checkpoint-72/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-72/model.safetensors b/runs/l2r90-baseline/checkpoint-72/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fe49a450132d2f620e9b8e49b23a55ec5b7271dc --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52490a039defecc0537c3f91d26e2ce9ddd6d0c081d6938213e70c85becb49de +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-72/optimizer.pt b/runs/l2r90-baseline/checkpoint-72/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..fa6f354bd02413d44b5a9e3adaa2228116294afc --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3eb623ba0676ffcdb34a9eb85048dcc4763cd55f4477c3b7b570e58e4079c863 +size 1166825338 diff --git a/runs/l2r90-baseline/checkpoint-72/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-72/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..5631561581b3634362004f6c036145ee8e54d0f4 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b184f8fc9facd6e4a6513602952b43b9446784300c3ea2da72aefb029f93aac1 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-72/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-72/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..5c9399d17854873e90933eddfa9944d5e3d0b38f --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:937f07b3f415f537fbc809c70f15f64b8410cd6a07adbe7d246f5ca7359b3b49 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-72/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-72/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..8a986c37b53237a7bde101804a5602d204f737ba --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07e3f7eacec877950b3dad99b106bf80da30203dda6159b06f954c15b7f5d367 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-72/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-72/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..3ec45d94b0661de4e43586eec83b2caf40c79a8e --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a5087a11a5f0a226f1607ec68740ea6baa2599559138de38da40f89dd666683 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-72/scheduler.pt b/runs/l2r90-baseline/checkpoint-72/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..43cf114590f1088fd438450226c42885d843da41 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:878ec53f5c2a524fc37af578b94a9025923faefea55f621be8e5643a79915d0f +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-72/tokenizer.json b/runs/l2r90-baseline/checkpoint-72/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-72/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-72/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-72/trainer_state.json b/runs/l2r90-baseline/checkpoint-72/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..37e1d12727703363c9e03304966f12832b9a9ffb --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/trainer_state.json @@ -0,0 +1,538 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 72, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.069613361332224e+16, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-72/training_args.bin b/runs/l2r90-baseline/checkpoint-72/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-72/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856 diff --git a/runs/l2r90-baseline/checkpoint-720/chat_template.jinja b/runs/l2r90-baseline/checkpoint-720/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r90-baseline/checkpoint-720/config.json b/runs/l2r90-baseline/checkpoint-720/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/l2r90-baseline/checkpoint-720/generation_config.json b/runs/l2r90-baseline/checkpoint-720/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r90-baseline/checkpoint-720/model.safetensors b/runs/l2r90-baseline/checkpoint-720/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f79755d77882de972fec84f50946b1190fb0ef7 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:180e7d4278088669ee7f57d6a61aaac74684a8f4ff9a36b69aac4e203a960731 +size 583356232 diff --git a/runs/l2r90-baseline/checkpoint-720/optimizer.pt b/runs/l2r90-baseline/checkpoint-720/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..ae1fd7129ce4b8110cee25739fcf91603e8e784b --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3173a961075e4fff1e113d04b690bfda0179968cf320f4766a35ed5d96c476b +size 1166825338 diff --git a/runs/l2r90-baseline/checkpoint-720/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-720/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..0977c0bf00c01a46d716db730a3c9bcc82905aaf --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a02d728a763af949adfec4bf1d121e1daf1ef655fd3b17a42c2399e0221768dd +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-720/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-720/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..4e5cd4f14d76f3fe2c0762093e6bb01d818def44 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d60087aa7d44f5751e6a709354a90b705404a21a9437680f20377cf12743341b +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-720/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-720/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..1dbfc9354146f87ee95d6d46bdc97b42ae7b6f65 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c962bfcedafeb4fe574f63e648405b9b28a01333d49c391529b2c7a4fead463 +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-720/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-720/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..0a96d4dec11fa27b14ffd309db2393d72f178feb --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:478ec022631f64178ab44a8355b560b75c942f4cf348cc636a8cb37ec3bec6cc +size 15024 diff --git a/runs/l2r90-baseline/checkpoint-720/scheduler.pt b/runs/l2r90-baseline/checkpoint-720/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..53133f8d1ca0990636972b84579aec1212febb2a --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea248f5cdc3b9f2d6793617acdaa11d93bab34c0a87ea94ea6aaabc148225da6 +size 1064 diff --git a/runs/l2r90-baseline/checkpoint-720/tokenizer.json b/runs/l2r90-baseline/checkpoint-720/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/l2r90-baseline/checkpoint-720/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-720/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r90-baseline/checkpoint-720/trainer_state.json b/runs/l2r90-baseline/checkpoint-720/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..0d47dcf6191b4c779498a357f56ba06e8b8c1324 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/trainer_state.json @@ -0,0 +1,5146 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 10.0, + "eval_steps": 500, + "global_step": 720, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.15997567772865295, + "learning_rate": 0.0, + "loss": 12.023977279663086, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.16017116606235504, + "learning_rate": 5.999999999999999e-06, + "loss": 12.022809982299805, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.1574592888355255, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.986297607421875, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.15066885948181152, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.91261100769043, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.14228598773479462, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.817420959472656, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.1410660743713379, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.70869255065918, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13418471813201904, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.597742080688477, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.12223710119724274, + "learning_rate": 4.2e-05, + "loss": 11.500282287597656, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11431518942117691, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.406240463256836, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11069707572460175, + "learning_rate": 5.399999999999999e-05, + "loss": 11.324350357055664, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10802163928747177, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.263275146484375, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10728747397661209, + "learning_rate": 6.599999999999999e-05, + "loss": 11.206270217895508, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10647646337747574, + "learning_rate": 7.199999999999999e-05, + "loss": 11.158624649047852, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10614501684904099, + "learning_rate": 7.8e-05, + "loss": 11.114002227783203, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10632237792015076, + "learning_rate": 8.4e-05, + "loss": 11.067373275756836, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.1061672493815422, + "learning_rate": 8.999999999999999e-05, + "loss": 11.015380859375, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10657714307308197, + "learning_rate": 9.599999999999999e-05, + "loss": 10.958027839660645, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10590068250894547, + "learning_rate": 0.000102, + "loss": 10.900931358337402, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591627657413483, + "learning_rate": 0.00010799999999999998, + "loss": 10.8344144821167, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10607340931892395, + "learning_rate": 0.00011399999999999999, + "loss": 10.761369705200195, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10447441786527634, + "learning_rate": 0.00011999999999999999, + "loss": 10.6970796585083, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.10463859140872955, + "learning_rate": 0.00012599999999999997, + "loss": 10.612152099609375, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.1049257144331932, + "learning_rate": 0.00013199999999999998, + "loss": 10.5253324508667, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10448617488145828, + "learning_rate": 0.000138, + "loss": 10.441542625427246, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10413316637277603, + "learning_rate": 0.00014399999999999998, + "loss": 10.353134155273438, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10503070801496506, + "learning_rate": 0.00015, + "loss": 10.251943588256836, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.10336724668741226, + "learning_rate": 0.000156, + "loss": 10.17003059387207, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10526089370250702, + "learning_rate": 0.000162, + "loss": 10.058032989501953, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.10343154519796371, + "learning_rate": 0.000168, + "loss": 9.968986511230469, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.10399148613214493, + "learning_rate": 0.00017399999999999997, + "loss": 9.858746528625488, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.1026586964726448, + "learning_rate": 0.00017999999999999998, + "loss": 9.766483306884766, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10253847390413284, + "learning_rate": 0.000186, + "loss": 9.65650463104248, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10166377574205399, + "learning_rate": 0.00019199999999999998, + "loss": 9.547554016113281, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10167674720287323, + "learning_rate": 0.000198, + "loss": 9.438922882080078, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09929938614368439, + "learning_rate": 0.000204, + "loss": 9.350300788879395, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09858057647943497, + "learning_rate": 0.00020999999999999998, + "loss": 9.234173774719238, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09838391095399857, + "learning_rate": 0.00021599999999999996, + "loss": 9.121601104736328, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.0974399745464325, + "learning_rate": 0.00022199999999999998, + "loss": 9.014242172241211, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09716248512268066, + "learning_rate": 0.00022799999999999999, + "loss": 8.907164573669434, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09500741958618164, + "learning_rate": 0.000234, + "loss": 8.81848430633545, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09266163408756256, + "learning_rate": 0.00023999999999999998, + "loss": 8.729146003723145, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.0910920575261116, + "learning_rate": 0.00024599999999999996, + "loss": 8.625696182250977, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.0894983559846878, + "learning_rate": 0.00025199999999999995, + "loss": 8.51424503326416, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08491404354572296, + "learning_rate": 0.000258, + "loss": 8.461417198181152, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.0824626088142395, + "learning_rate": 0.00026399999999999997, + "loss": 8.371938705444336, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.0831412672996521, + "learning_rate": 0.00027, + "loss": 8.28670597076416, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07819175720214844, + "learning_rate": 0.000276, + "loss": 8.186386108398438, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07501859962940216, + "learning_rate": 0.00028199999999999997, + "loss": 8.095057487487793, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0688866451382637, + "learning_rate": 0.00028799999999999995, + "loss": 8.032234191894531, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06329295784235, + "learning_rate": 0.000294, + "loss": 7.964580059051514, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.06046910956501961, + "learning_rate": 0.0003, + "loss": 7.915485858917236, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.050732146948575974, + "learning_rate": 0.00030599999999999996, + "loss": 7.86663818359375, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04609096050262451, + "learning_rate": 0.000312, + "loss": 7.804510116577148, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03886866196990013, + "learning_rate": 0.000318, + "loss": 7.783796310424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.034636761993169785, + "learning_rate": 0.000324, + "loss": 7.746722221374512, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.02844400331377983, + "learning_rate": 0.00033, + "loss": 7.712285041809082, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.026054631918668747, + "learning_rate": 0.000336, + "loss": 7.708642482757568, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.0225310567766428, + "learning_rate": 0.00034199999999999996, + "loss": 7.6868672370910645, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.018994266167283058, + "learning_rate": 0.00034799999999999995, + "loss": 7.6628804206848145, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017860861495137215, + "learning_rate": 0.00035399999999999993, + "loss": 7.652166366577148, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.019245322793722153, + "learning_rate": 0.00035999999999999997, + "loss": 7.612408638000488, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.01606360450387001, + "learning_rate": 0.00036599999999999995, + "loss": 7.627114295959473, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.01609027571976185, + "learning_rate": 0.000372, + "loss": 7.614614486694336, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.014387412928044796, + "learning_rate": 0.00037799999999999997, + "loss": 7.593182563781738, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.016893157735466957, + "learning_rate": 0.00038399999999999996, + "loss": 7.60258674621582, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01590251922607422, + "learning_rate": 0.00039, + "loss": 7.590658187866211, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.012655598111450672, + "learning_rate": 0.000396, + "loss": 7.571450233459473, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.013261383399367332, + "learning_rate": 0.000402, + "loss": 7.5402021408081055, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.01822199858725071, + "learning_rate": 0.000408, + "loss": 7.541982650756836, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.0237593874335289, + "learning_rate": 0.0004139999999999999, + "loss": 7.508282661437988, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.05531860142946243, + "learning_rate": 0.00041999999999999996, + "loss": 7.518495082855225, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.12719468772411346, + "learning_rate": 0.00042599999999999995, + "loss": 7.575924873352051, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.544726848602295, + "eval_runtime": 56.6032, + "eval_samples_per_second": 43.142, + "eval_steps_per_second": 1.36, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.06889434158802032, + "learning_rate": 0.00043199999999999993, + "loss": 7.532279968261719, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.043669551610946655, + "learning_rate": 0.00043799999999999997, + "loss": 7.538493633270264, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.02140442468225956, + "learning_rate": 0.00044399999999999995, + "loss": 7.478395462036133, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.03172273188829422, + "learning_rate": 0.00045, + "loss": 7.495583534240723, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.04667133465409279, + "learning_rate": 0.00045599999999999997, + "loss": 7.470980644226074, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.05474933981895447, + "learning_rate": 0.00046199999999999995, + "loss": 7.445516109466553, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01863257959485054, + "learning_rate": 0.000468, + "loss": 7.463066101074219, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.037733182311058044, + "learning_rate": 0.000474, + "loss": 7.457036018371582, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.026289211586117744, + "learning_rate": 0.00047999999999999996, + "loss": 7.440082550048828, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.027869857847690582, + "learning_rate": 0.000486, + "loss": 7.412266731262207, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.021711770445108414, + "learning_rate": 0.0004919999999999999, + "loss": 7.384769439697266, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.03738119453191757, + "learning_rate": 0.000498, + "loss": 7.355679512023926, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.034208305180072784, + "learning_rate": 0.0005039999999999999, + "loss": 7.343960285186768, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.020337563008069992, + "learning_rate": 0.0005099999999999999, + "loss": 7.337154388427734, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.036635156720876694, + "learning_rate": 0.000516, + "loss": 7.334864139556885, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.0275711789727211, + "learning_rate": 0.000522, + "loss": 7.3356451988220215, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.022682661190629005, + "learning_rate": 0.0005279999999999999, + "loss": 7.28330659866333, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.034273840487003326, + "learning_rate": 0.000534, + "loss": 7.287396430969238, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.02616060897707939, + "learning_rate": 0.00054, + "loss": 7.272294044494629, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.01899445243179798, + "learning_rate": 0.0005459999999999999, + "loss": 7.239999771118164, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.020839497447013855, + "learning_rate": 0.000552, + "loss": 7.235346794128418, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.023729125037789345, + "learning_rate": 0.000558, + "loss": 7.1834821701049805, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.0318775475025177, + "learning_rate": 0.0005639999999999999, + "loss": 7.198540687561035, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.015114367939531803, + "learning_rate": 0.00057, + "loss": 7.212254524230957, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.02599942311644554, + "learning_rate": 0.0005759999999999999, + "loss": 7.164607048034668, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.029708007350564003, + "learning_rate": 0.0005819999999999999, + "loss": 7.175436019897461, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.0463266558945179, + "learning_rate": 0.000588, + "loss": 7.173397541046143, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.06682766228914261, + "learning_rate": 0.0005939999999999999, + "loss": 7.164839744567871, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.07400168478488922, + "learning_rate": 0.0006, + "loss": 7.173515319824219, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.04783684387803078, + "learning_rate": 0.0005999998803562321, + "loss": 7.16644287109375, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.042767491191625595, + "learning_rate": 0.0005999995214250253, + "loss": 7.145862579345703, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.028570281341671944, + "learning_rate": 0.0005999989232066684, + "loss": 7.101897716522217, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.04605371132493019, + "learning_rate": 0.0005999980857016436, + "loss": 7.111356735229492, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.02357972227036953, + "learning_rate": 0.0005999970089106256, + "loss": 7.06209659576416, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.04726218059659004, + "learning_rate": 0.0005999956928344818, + "loss": 7.099649429321289, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.02519933320581913, + "learning_rate": 0.0005999941374742726, + "loss": 7.050154209136963, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.027398832142353058, + "learning_rate": 0.0005999923428312514, + "loss": 7.028330326080322, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.027254166081547737, + "learning_rate": 0.0005999903089068637, + "loss": 7.049069404602051, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.02054189145565033, + "learning_rate": 0.0005999880357027487, + "loss": 7.0402045249938965, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.021891873329877853, + "learning_rate": 0.0005999855232207374, + "loss": 7.023792743682861, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.017824960872530937, + "learning_rate": 0.0005999827714628542, + "loss": 7.002220153808594, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.022561997175216675, + "learning_rate": 0.0005999797804313163, + "loss": 7.0106201171875, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.016903111711144447, + "learning_rate": 0.0005999765501285333, + "loss": 6.970179557800293, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.02128630317747593, + "learning_rate": 0.000599973080557108, + "loss": 7.003076553344727, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.02123585157096386, + "learning_rate": 0.0005999693717198356, + "loss": 6.9614105224609375, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.01661432348191738, + "learning_rate": 0.0005999654236197044, + "loss": 6.955131530761719, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.01780594512820244, + "learning_rate": 0.0005999612362598951, + "loss": 6.93811559677124, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.020178694278001785, + "learning_rate": 0.0005999568096437816, + "loss": 6.902839660644531, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.033332180231809616, + "learning_rate": 0.0005999521437749303, + "loss": 6.918459415435791, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.044249728322029114, + "learning_rate": 0.0005999472386571002, + "loss": 6.946916103363037, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.04895849898457527, + "learning_rate": 0.0005999420942942435, + "loss": 6.9353156089782715, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.024643104523420334, + "learning_rate": 0.000599936710690505, + "loss": 6.915379524230957, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.02990383468568325, + "learning_rate": 0.0005999310878502218, + "loss": 6.898656845092773, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.033634357154369354, + "learning_rate": 0.0005999252257779244, + "loss": 6.923366546630859, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.03218965232372284, + "learning_rate": 0.0005999191244783357, + "loss": 6.923885345458984, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.02356622740626335, + "learning_rate": 0.0005999127839563715, + "loss": 6.899965286254883, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.03445356339216232, + "learning_rate": 0.00059990620421714, + "loss": 6.883448600769043, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.028274044394493103, + "learning_rate": 0.0005998993852659426, + "loss": 6.884991645812988, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.017005041241645813, + "learning_rate": 0.0005998923271082733, + "loss": 6.8683953285217285, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.031443849205970764, + "learning_rate": 0.0005998850297498183, + "loss": 6.832046031951904, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.029097655788064003, + "learning_rate": 0.0005998774931964574, + "loss": 6.835471153259277, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07707645744085312, + "learning_rate": 0.0005998697174542624, + "loss": 6.857353210449219, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.09299153089523315, + "learning_rate": 0.0005998617025294983, + "loss": 6.933018207550049, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047855719923973083, + "learning_rate": 0.0005998534484286223, + "loss": 6.8583083152771, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.07969525456428528, + "learning_rate": 0.0005998449551582847, + "loss": 6.864906311035156, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.1365756392478943, + "learning_rate": 0.0005998362227253285, + "loss": 6.903354644775391, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.1410154104232788, + "learning_rate": 0.0005998272511367889, + "loss": 6.947896957397461, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.07029050588607788, + "learning_rate": 0.0005998180403998947, + "loss": 6.910849571228027, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.06964805722236633, + "learning_rate": 0.0005998085905220664, + "loss": 6.882598876953125, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.05349893495440483, + "learning_rate": 0.0005997989015109177, + "loss": 6.836871147155762, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.05424615368247032, + "learning_rate": 0.000599788973374255, + "loss": 6.832496643066406, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.044589463621377945, + "learning_rate": 0.000599778806120077, + "loss": 6.8051371574401855, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.840139865875244, + "eval_runtime": 56.8976, + "eval_samples_per_second": 42.919, + "eval_steps_per_second": 1.353, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.04068000614643097, + "learning_rate": 0.0005997683997565754, + "loss": 6.811692714691162, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03961706906557083, + "learning_rate": 0.0005997577542921344, + "loss": 6.788667678833008, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03070618025958538, + "learning_rate": 0.0005997468697353309, + "loss": 6.7994585037231445, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.036584168672561646, + "learning_rate": 0.0005997357460949342, + "loss": 6.767686367034912, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.02327805571258068, + "learning_rate": 0.0005997243833799068, + "loss": 6.784345626831055, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.028179455548524857, + "learning_rate": 0.000599712781599403, + "loss": 6.769927024841309, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.025651710107922554, + "learning_rate": 0.0005997009407627704, + "loss": 6.7423295974731445, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.025366952642798424, + "learning_rate": 0.000599688860879549, + "loss": 6.736351013183594, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.023156974464654922, + "learning_rate": 0.0005996765419594711, + "loss": 6.697917461395264, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.021937215700745583, + "learning_rate": 0.000599663984012462, + "loss": 6.687577247619629, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.020132191479206085, + "learning_rate": 0.0005996511870486393, + "loss": 6.69163179397583, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.018260158598423004, + "learning_rate": 0.0005996381510783135, + "loss": 6.673725128173828, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.017023146152496338, + "learning_rate": 0.0005996248761119872, + "loss": 6.660531520843506, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.020211776718497276, + "learning_rate": 0.0005996113621603559, + "loss": 6.6298699378967285, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.01748521998524666, + "learning_rate": 0.0005995976092343075, + "loss": 6.626924514770508, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.023990269750356674, + "learning_rate": 0.0005995836173449227, + "loss": 6.628607749938965, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.060859110206365585, + "learning_rate": 0.0005995693865034743, + "loss": 6.625316619873047, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.12787196040153503, + "learning_rate": 0.0005995549167214278, + "loss": 6.758873462677002, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.0820903554558754, + "learning_rate": 0.0005995402080104414, + "loss": 6.788875579833984, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.08051387965679169, + "learning_rate": 0.0005995252603823656, + "loss": 6.788204193115234, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.04531198740005493, + "learning_rate": 0.0005995100738492433, + "loss": 6.734226703643799, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03978786990046501, + "learning_rate": 0.0005994946484233102, + "loss": 6.685703277587891, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.0414281040430069, + "learning_rate": 0.0005994789841169941, + "loss": 6.713569641113281, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.040948331356048584, + "learning_rate": 0.0005994630809429156, + "loss": 6.695756435394287, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.04339255020022392, + "learning_rate": 0.0005994469389138875, + "loss": 6.697275638580322, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.038956254720687866, + "learning_rate": 0.0005994305580429152, + "loss": 6.6156110763549805, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.03434654325246811, + "learning_rate": 0.0005994139383431966, + "loss": 6.639013290405273, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.0459795817732811, + "learning_rate": 0.0005993970798281218, + "loss": 6.603760242462158, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.032164230942726135, + "learning_rate": 0.000599379982511273, + "loss": 6.6016459465026855, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.02468927390873432, + "learning_rate": 0.0005993626464064259, + "loss": 6.589484214782715, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.02615884505212307, + "learning_rate": 0.0005993450715275474, + "loss": 6.5663275718688965, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.02823791094124317, + "learning_rate": 0.0005993272578887974, + "loss": 6.585080623626709, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.02933267503976822, + "learning_rate": 0.000599309205504528, + "loss": 6.549171447753906, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.020691825076937675, + "learning_rate": 0.0005992909143892837, + "loss": 6.518960475921631, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.02271026186645031, + "learning_rate": 0.0005992723845578016, + "loss": 6.519137382507324, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.028786325827240944, + "learning_rate": 0.0005992536160250103, + "loss": 6.516097545623779, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.024719521403312683, + "learning_rate": 0.0005992346088060317, + "loss": 6.495189666748047, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.02592875435948372, + "learning_rate": 0.0005992153629161793, + "loss": 6.505522727966309, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.022884991019964218, + "learning_rate": 0.0005991958783709593, + "loss": 6.508190155029297, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.01747789792716503, + "learning_rate": 0.0005991761551860702, + "loss": 6.475917816162109, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.017053667455911636, + "learning_rate": 0.0005991561933774022, + "loss": 6.474056243896484, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.0169149748980999, + "learning_rate": 0.0005991359929610386, + "loss": 6.437000751495361, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.017995426431298256, + "learning_rate": 0.0005991155539532544, + "loss": 6.416827201843262, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.023745683953166008, + "learning_rate": 0.0005990948763705167, + "loss": 6.447346210479736, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01796909235417843, + "learning_rate": 0.0005990739602294852, + "loss": 6.426738739013672, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.014229115098714828, + "learning_rate": 0.0005990528055470118, + "loss": 6.4247660636901855, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.020987126976251602, + "learning_rate": 0.0005990314123401403, + "loss": 6.384764194488525, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.037671539932489395, + "learning_rate": 0.0005990097806261068, + "loss": 6.393845558166504, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.06533686071634293, + "learning_rate": 0.0005989879104223397, + "loss": 6.418580055236816, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.09679993242025375, + "learning_rate": 0.0005989658017464593, + "loss": 6.455891132354736, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.08761037141084671, + "learning_rate": 0.0005989434546162782, + "loss": 6.458295822143555, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.06243691220879555, + "learning_rate": 0.0005989208690498012, + "loss": 6.4045023918151855, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.052615419030189514, + "learning_rate": 0.0005988980450652248, + "loss": 6.393614292144775, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.04153689369559288, + "learning_rate": 0.0005988749826809381, + "loss": 6.329081058502197, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.03536545857787132, + "learning_rate": 0.0005988516819155222, + "loss": 6.382253170013428, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.03506210446357727, + "learning_rate": 0.0005988281427877498, + "loss": 6.3689374923706055, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.028134305030107498, + "learning_rate": 0.000598804365316586, + "loss": 6.350457191467285, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.025324443355202675, + "learning_rate": 0.0005987803495211879, + "loss": 6.3551506996154785, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.027925321832299232, + "learning_rate": 0.0005987560954209047, + "loss": 6.342905044555664, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.02472207508981228, + "learning_rate": 0.0005987316030352773, + "loss": 6.320896148681641, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.022138327360153198, + "learning_rate": 0.000598706872384039, + "loss": 6.326817512512207, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.02824472263455391, + "learning_rate": 0.0005986819034871147, + "loss": 6.333435535430908, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.015822919085621834, + "learning_rate": 0.0005986566963646214, + "loss": 6.287126541137695, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.02211311087012291, + "learning_rate": 0.000598631251036868, + "loss": 6.274896144866943, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.01805170439183712, + "learning_rate": 0.0005986055675243555, + "loss": 6.27794075012207, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.02028656005859375, + "learning_rate": 0.0005985796458477765, + "loss": 6.288621425628662, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02157517708837986, + "learning_rate": 0.0005985534860280155, + "loss": 6.297999382019043, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.028337396681308746, + "learning_rate": 0.0005985270880861493, + "loss": 6.256705284118652, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.036196526139974594, + "learning_rate": 0.0005985004520434462, + "loss": 6.259498596191406, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.04559934884309769, + "learning_rate": 0.000598473577921366, + "loss": 6.26982307434082, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.05263037979602814, + "learning_rate": 0.0005984464657415611, + "loss": 6.276828765869141, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.046778593212366104, + "learning_rate": 0.0005984191155258751, + "loss": 6.2884979248046875, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.268282413482666, + "eval_runtime": 57.2037, + "eval_samples_per_second": 42.69, + "eval_steps_per_second": 1.346, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.029683491215109825, + "learning_rate": 0.0005983915272963436, + "loss": 6.2267165184021, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.03019959293305874, + "learning_rate": 0.0005983637010751941, + "loss": 6.223328590393066, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.031647924333810806, + "learning_rate": 0.0005983356368848455, + "loss": 6.228819370269775, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.025646064430475235, + "learning_rate": 0.0005983073347479086, + "loss": 6.219881534576416, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03026074543595314, + "learning_rate": 0.000598278794687186, + "loss": 6.172661304473877, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.027252312749624252, + "learning_rate": 0.0005982500167256719, + "loss": 6.181373119354248, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03526923805475235, + "learning_rate": 0.0005982210008865522, + "loss": 6.194186210632324, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03685254603624344, + "learning_rate": 0.0005981917471932045, + "loss": 6.187342166900635, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.036119766533374786, + "learning_rate": 0.0005981622556691978, + "loss": 6.200807571411133, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.036502398550510406, + "learning_rate": 0.0005981325263382931, + "loss": 6.162562370300293, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.03517023101449013, + "learning_rate": 0.0005981025592244425, + "loss": 6.149142265319824, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.023205652832984924, + "learning_rate": 0.00059807235435179, + "loss": 6.148383140563965, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.0221528522670269, + "learning_rate": 0.0005980419117446715, + "loss": 6.143627643585205, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.024407457560300827, + "learning_rate": 0.0005980112314276138, + "loss": 6.146297931671143, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.02728901617228985, + "learning_rate": 0.0005979803134253354, + "loss": 6.114583969116211, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.039508312940597534, + "learning_rate": 0.0005979491577627464, + "loss": 6.149223327636719, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.048040807247161865, + "learning_rate": 0.0005979177644649485, + "loss": 6.122624397277832, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.056522514671087265, + "learning_rate": 0.0005978861335572346, + "loss": 6.141239166259766, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.06174373999238014, + "learning_rate": 0.0005978542650650892, + "loss": 6.152597427368164, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.07200063765048981, + "learning_rate": 0.0005978221590141881, + "loss": 6.107697486877441, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.1010260134935379, + "learning_rate": 0.0005977898154303987, + "loss": 6.191191673278809, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.06763981282711029, + "learning_rate": 0.0005977572343397794, + "loss": 6.132532119750977, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.042127929627895355, + "learning_rate": 0.0005977244157685805, + "loss": 6.165022850036621, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.03653500974178314, + "learning_rate": 0.000597691359743243, + "loss": 6.143064975738525, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.03675027936697006, + "learning_rate": 0.0005976580662903999, + "loss": 6.108728408813477, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.03305647522211075, + "learning_rate": 0.0005976245354368749, + "loss": 6.108780860900879, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.026183469220995903, + "learning_rate": 0.0005975907672096832, + "loss": 6.088522911071777, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.02158971130847931, + "learning_rate": 0.0005975567616360313, + "loss": 6.063396453857422, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.029356449842453003, + "learning_rate": 0.0005975225187433169, + "loss": 6.10193395614624, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.020083140581846237, + "learning_rate": 0.000597488038559129, + "loss": 6.059609413146973, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02572067640721798, + "learning_rate": 0.0005974533211112474, + "loss": 6.08794641494751, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.02591501921415329, + "learning_rate": 0.0005974183664276434, + "loss": 6.081871032714844, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.023668579757213593, + "learning_rate": 0.0005973831745364795, + "loss": 6.077000617980957, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.02440894953906536, + "learning_rate": 0.000597347745466109, + "loss": 6.063971519470215, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.024253705516457558, + "learning_rate": 0.0005973120792450766, + "loss": 6.08235502243042, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.024689998477697372, + "learning_rate": 0.0005972761759021178, + "loss": 5.998595237731934, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.028521878644824028, + "learning_rate": 0.0005972400354661594, + "loss": 5.979642868041992, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.030266916379332542, + "learning_rate": 0.000597203657966319, + "loss": 6.000593185424805, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.033342741429805756, + "learning_rate": 0.0005971670434319053, + "loss": 6.004687309265137, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03348678722977638, + "learning_rate": 0.0005971301918924182, + "loss": 6.029177665710449, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.025689199566841125, + "learning_rate": 0.0005970931033775479, + "loss": 5.98867130279541, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.027882622554898262, + "learning_rate": 0.0005970557779171763, + "loss": 5.974981784820557, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04662966728210449, + "learning_rate": 0.0005970182155413756, + "loss": 5.979356288909912, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.07830756902694702, + "learning_rate": 0.0005969804162804093, + "loss": 6.011981964111328, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.07885197550058365, + "learning_rate": 0.0005969423801647314, + "loss": 6.019845008850098, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.0604281947016716, + "learning_rate": 0.0005969041072249872, + "loss": 5.991020679473877, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.059781573712825775, + "learning_rate": 0.0005968655974920124, + "loss": 5.999235153198242, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.04918011277914047, + "learning_rate": 0.0005968268509968335, + "loss": 6.012720108032227, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.03718101978302002, + "learning_rate": 0.0005967878677706678, + "loss": 5.986680507659912, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.03947242721915245, + "learning_rate": 0.0005967486478449236, + "loss": 5.959175109863281, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.04288070648908615, + "learning_rate": 0.0005967091912511996, + "loss": 5.976179122924805, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.03525253012776375, + "learning_rate": 0.0005966694980212851, + "loss": 5.987687587738037, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.030190549790859222, + "learning_rate": 0.0005966295681871604, + "loss": 5.965185642242432, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.027395015582442284, + "learning_rate": 0.0005965894017809962, + "loss": 5.958373069763184, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02688959613442421, + "learning_rate": 0.0005965489988351539, + "loss": 5.959633827209473, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.02119498886168003, + "learning_rate": 0.0005965083593821853, + "loss": 5.934293270111084, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022015074267983437, + "learning_rate": 0.000596467483454833, + "loss": 5.914854049682617, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.021204788237810135, + "learning_rate": 0.0005964263710860299, + "loss": 5.900177955627441, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.018889140337705612, + "learning_rate": 0.0005963850223088995, + "loss": 5.873917579650879, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.018911773338913918, + "learning_rate": 0.0005963434371567559, + "loss": 5.88484001159668, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.019570615142583847, + "learning_rate": 0.0005963016156631033, + "loss": 5.886139392852783, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.02666759118437767, + "learning_rate": 0.0005962595578616367, + "loss": 5.876997470855713, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.03528648987412453, + "learning_rate": 0.0005962172637862411, + "loss": 5.8770318031311035, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.0549609512090683, + "learning_rate": 0.0005961747334709924, + "loss": 5.86689567565918, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.07202128320932388, + "learning_rate": 0.0005961319669501562, + "loss": 5.897315979003906, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.04940229654312134, + "learning_rate": 0.0005960889642581887, + "loss": 5.914303779602051, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.05204314738512039, + "learning_rate": 0.0005960457254297367, + "loss": 5.918832302093506, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.0608876571059227, + "learning_rate": 0.0005960022504996369, + "loss": 5.898425579071045, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.0627935454249382, + "learning_rate": 0.0005959585395029158, + "loss": 5.890061378479004, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.04895327612757683, + "learning_rate": 0.000595914592474791, + "loss": 5.885444164276123, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.04046280309557915, + "learning_rate": 0.0005958704094506698, + "loss": 5.8441972732543945, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.045558422803878784, + "learning_rate": 0.0005958259904661497, + "loss": 5.856559753417969, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 5.899378776550293, + "eval_runtime": 58.627, + "eval_samples_per_second": 41.653, + "eval_steps_per_second": 1.313, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04504650831222534, + "learning_rate": 0.0005957813355570178, + "loss": 5.841728210449219, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.036045704036951065, + "learning_rate": 0.0005957364447592524, + "loss": 5.8451619148254395, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.029153067618608475, + "learning_rate": 0.0005956913181090208, + "loss": 5.811412811279297, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.027814218774437904, + "learning_rate": 0.0005956459556426809, + "loss": 5.848013877868652, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.025015871971845627, + "learning_rate": 0.0005956003573967802, + "loss": 5.781689643859863, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.02065589465200901, + "learning_rate": 0.0005955545234080567, + "loss": 5.783626079559326, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.023988133296370506, + "learning_rate": 0.0005955084537134378, + "loss": 5.800251483917236, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.01968221738934517, + "learning_rate": 0.0005954621483500411, + "loss": 5.7953410148620605, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.021222546696662903, + "learning_rate": 0.0005954156073551739, + "loss": 5.765391826629639, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.018830647692084312, + "learning_rate": 0.0005953688307663336, + "loss": 5.77407693862915, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.021495003253221512, + "learning_rate": 0.0005953218186212072, + "loss": 5.756778717041016, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02049877494573593, + "learning_rate": 0.0005952745709576714, + "loss": 5.745931625366211, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.020010637119412422, + "learning_rate": 0.000595227087813793, + "loss": 5.755594253540039, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.030872922390699387, + "learning_rate": 0.0005951793692278282, + "loss": 5.775322914123535, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.056580882519483566, + "learning_rate": 0.0005951314152382229, + "loss": 5.771434783935547, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.07649239897727966, + "learning_rate": 0.0005950832258836129, + "loss": 5.812885284423828, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.05227350816130638, + "learning_rate": 0.0005950348012028235, + "loss": 5.755342960357666, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0459996834397316, + "learning_rate": 0.0005949861412348694, + "loss": 5.770654201507568, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04197477549314499, + "learning_rate": 0.0005949372460189555, + "loss": 5.7795915603637695, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.039317045360803604, + "learning_rate": 0.0005948881155944753, + "loss": 5.7311177253723145, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.0334056094288826, + "learning_rate": 0.0005948387500010126, + "loss": 5.737773895263672, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.02994806133210659, + "learning_rate": 0.0005947891492783401, + "loss": 5.693323612213135, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.03594277426600456, + "learning_rate": 0.0005947393134664205, + "loss": 5.7090654373168945, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.03271142393350601, + "learning_rate": 0.0005946892426054054, + "loss": 5.691743850708008, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.030780376866459846, + "learning_rate": 0.0005946389367356361, + "loss": 5.72518253326416, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.02948964200913906, + "learning_rate": 0.0005945883958976432, + "loss": 5.725653648376465, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.024995718151330948, + "learning_rate": 0.0005945376201321464, + "loss": 5.695929527282715, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.026229677721858025, + "learning_rate": 0.0005944866094800548, + "loss": 5.717738628387451, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02641691267490387, + "learning_rate": 0.0005944353639824669, + "loss": 5.687970161437988, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.022786157205700874, + "learning_rate": 0.0005943838836806702, + "loss": 5.651585578918457, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.023804830387234688, + "learning_rate": 0.0005943321686161414, + "loss": 5.67464542388916, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.025330908596515656, + "learning_rate": 0.0005942802188305464, + "loss": 5.6631011962890625, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.03142685443162918, + "learning_rate": 0.0005942280343657403, + "loss": 5.651077747344971, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.04587113484740257, + "learning_rate": 0.0005941756152637671, + "loss": 5.647086143493652, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.04738520458340645, + "learning_rate": 0.00059412296156686, + "loss": 5.701745986938477, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.028803404420614243, + "learning_rate": 0.0005940700733174409, + "loss": 5.648317813873291, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.041121263056993484, + "learning_rate": 0.0005940169505581213, + "loss": 5.647163391113281, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.05275499075651169, + "learning_rate": 0.0005939635933317009, + "loss": 5.666481018066406, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.060963090509176254, + "learning_rate": 0.0005939100016811688, + "loss": 5.645685195922852, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.05624202638864517, + "learning_rate": 0.0005938561756497028, + "loss": 5.6632795333862305, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.07903645187616348, + "learning_rate": 0.0005938021152806696, + "loss": 5.693739891052246, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.06570398062467575, + "learning_rate": 0.0005937478206176248, + "loss": 5.675642013549805, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.05279144272208214, + "learning_rate": 0.0005936932917043124, + "loss": 5.685312747955322, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.05221335217356682, + "learning_rate": 0.0005936385285846651, + "loss": 5.665262222290039, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.044064246118068695, + "learning_rate": 0.0005935835313028052, + "loss": 5.647260665893555, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.03285257890820503, + "learning_rate": 0.0005935282999030426, + "loss": 5.631403923034668, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.03275465965270996, + "learning_rate": 0.0005934728344298763, + "loss": 5.6139068603515625, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.0317506417632103, + "learning_rate": 0.000593417134927994, + "loss": 5.6341753005981445, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.0239992905408144, + "learning_rate": 0.0005933612014422715, + "loss": 5.601463794708252, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.02451021410524845, + "learning_rate": 0.0005933050340177736, + "loss": 5.581960201263428, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.02237604930996895, + "learning_rate": 0.0005932486326997534, + "loss": 5.602834224700928, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.01903100311756134, + "learning_rate": 0.0005931919975336523, + "loss": 5.604010105133057, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.018574781715869904, + "learning_rate": 0.0005931351285651004, + "loss": 5.556118965148926, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.019580982625484467, + "learning_rate": 0.0005930780258399157, + "loss": 5.594083786010742, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.02127210609614849, + "learning_rate": 0.0005930206894041051, + "loss": 5.574706077575684, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.02134147845208645, + "learning_rate": 0.0005929631193038634, + "loss": 5.57991361618042, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.02030216157436371, + "learning_rate": 0.0005929053155855739, + "loss": 5.554903984069824, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01813247986137867, + "learning_rate": 0.0005928472782958079, + "loss": 5.550544261932373, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.021302592009305954, + "learning_rate": 0.000592789007481325, + "loss": 5.560070991516113, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.029217662289738655, + "learning_rate": 0.0005927305031890732, + "loss": 5.517913818359375, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.03720416501164436, + "learning_rate": 0.000592671765466188, + "loss": 5.551721096038818, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.03990121930837631, + "learning_rate": 0.0005926127943599934, + "loss": 5.5301032066345215, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.03713113069534302, + "learning_rate": 0.0005925535899180015, + "loss": 5.5560712814331055, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.039228230714797974, + "learning_rate": 0.0005924941521879122, + "loss": 5.531858444213867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.030399464070796967, + "learning_rate": 0.0005924344812176134, + "loss": 5.5186357498168945, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.0415034294128418, + "learning_rate": 0.0005923745770551808, + "loss": 5.512445449829102, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.04447956383228302, + "learning_rate": 0.0005923144397488782, + "loss": 5.491172790527344, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.03943555802106857, + "learning_rate": 0.0005922540693471572, + "loss": 5.51191520690918, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.03963116183876991, + "learning_rate": 0.0005921934658986571, + "loss": 5.513973236083984, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.04397537559270859, + "learning_rate": 0.000592132629452205, + "loss": 5.515983581542969, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.04135015606880188, + "learning_rate": 0.0005920715600568156, + "loss": 5.465071201324463, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.03931731730699539, + "learning_rate": 0.0005920102577616916, + "loss": 5.504716873168945, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 5.549652099609375, + "eval_runtime": 57.761, + "eval_samples_per_second": 42.278, + "eval_steps_per_second": 1.333, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.03958690166473389, + "learning_rate": 0.0005919487226162231, + "loss": 5.459286689758301, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.03299890458583832, + "learning_rate": 0.0005918869546699877, + "loss": 5.45464563369751, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.028040863573551178, + "learning_rate": 0.0005918249539727509, + "loss": 5.444796562194824, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.028365720063447952, + "learning_rate": 0.0005917627205744655, + "loss": 5.447852611541748, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.02822224237024784, + "learning_rate": 0.0005917002545252719, + "loss": 5.453883171081543, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.036075398325920105, + "learning_rate": 0.0005916375558754977, + "loss": 5.462671279907227, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04683876410126686, + "learning_rate": 0.0005915746246756581, + "loss": 5.443917274475098, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.05457588657736778, + "learning_rate": 0.0005915114609764558, + "loss": 5.444374084472656, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.04250836744904518, + "learning_rate": 0.0005914480648287804, + "loss": 5.435029983520508, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.029477844014763832, + "learning_rate": 0.0005913844362837093, + "loss": 5.421862602233887, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.03673284500837326, + "learning_rate": 0.0005913205753925066, + "loss": 5.430391311645508, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.030489861965179443, + "learning_rate": 0.0005912564822066241, + "loss": 5.389314651489258, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.04037046805024147, + "learning_rate": 0.0005911921567777005, + "loss": 5.425060272216797, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.0445217490196228, + "learning_rate": 0.0005911275991575613, + "loss": 5.437565326690674, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.03885014355182648, + "learning_rate": 0.0005910628093982198, + "loss": 5.410930633544922, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.0334964245557785, + "learning_rate": 0.0005909977875518759, + "loss": 5.412092208862305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.03499286621809006, + "learning_rate": 0.0005909325336709164, + "loss": 5.382216453552246, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.04171644523739815, + "learning_rate": 0.0005908670478079152, + "loss": 5.387385368347168, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.057558078318834305, + "learning_rate": 0.000590801330015633, + "loss": 5.437925815582275, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.053846780210733414, + "learning_rate": 0.0005907353803470177, + "loss": 5.41644811630249, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.027259036898612976, + "learning_rate": 0.0005906691988552034, + "loss": 5.409185409545898, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.03394043818116188, + "learning_rate": 0.0005906027855935115, + "loss": 5.377973556518555, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.03731574863195419, + "learning_rate": 0.0005905361406154501, + "loss": 5.382505893707275, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.04705141484737396, + "learning_rate": 0.0005904692639747137, + "loss": 5.413661003112793, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.044811323285102844, + "learning_rate": 0.0005904021557251837, + "loss": 5.417835235595703, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.04000752419233322, + "learning_rate": 0.0005903348159209277, + "loss": 5.406160354614258, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.038364164531230927, + "learning_rate": 0.0005902672446162007, + "loss": 5.357950687408447, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.030500127002596855, + "learning_rate": 0.0005901994418654432, + "loss": 5.344951629638672, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.030644729733467102, + "learning_rate": 0.0005901314077232829, + "loss": 5.3599348068237305, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.037976790219545364, + "learning_rate": 0.0005900631422445335, + "loss": 5.354084491729736, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.04106011986732483, + "learning_rate": 0.0005899946454841955, + "loss": 5.363678455352783, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.04334832355380058, + "learning_rate": 0.0005899259174974553, + "loss": 5.350959300994873, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.049939028918743134, + "learning_rate": 0.000589856958339686, + "loss": 5.3317155838012695, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.05300058051943779, + "learning_rate": 0.0005897877680664465, + "loss": 5.353314399719238, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.05943184345960617, + "learning_rate": 0.0005897183467334822, + "loss": 5.3683929443359375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.05600789934396744, + "learning_rate": 0.0005896486943967249, + "loss": 5.345516204833984, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.05392718315124512, + "learning_rate": 0.0005895788111122919, + "loss": 5.291389465332031, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.037913475185632706, + "learning_rate": 0.0005895086969364871, + "loss": 5.353423118591309, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.03362123295664787, + "learning_rate": 0.0005894383519258001, + "loss": 5.379500389099121, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032558735460042953, + "learning_rate": 0.0005893677761369066, + "loss": 5.34043025970459, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.04074402526021004, + "learning_rate": 0.0005892969696266683, + "loss": 5.303869247436523, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.04270484298467636, + "learning_rate": 0.0005892259324521328, + "loss": 5.27297830581665, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.04604959115386009, + "learning_rate": 0.0005891546646705334, + "loss": 5.313268184661865, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.04462246224284172, + "learning_rate": 0.0005890831663392891, + "loss": 5.31167459487915, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.04966094344854355, + "learning_rate": 0.0005890114375160052, + "loss": 5.316695690155029, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.04662073776125908, + "learning_rate": 0.0005889394782584718, + "loss": 5.316898822784424, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.04380315542221069, + "learning_rate": 0.0005888672886246656, + "loss": 5.285087585449219, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03239579498767853, + "learning_rate": 0.0005887948686727483, + "loss": 5.304283142089844, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.029690183699131012, + "learning_rate": 0.0005887222184610675, + "loss": 5.305715084075928, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.03444967418909073, + "learning_rate": 0.0005886493380481559, + "loss": 5.2707672119140625, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.03302565962076187, + "learning_rate": 0.0005885762274927322, + "loss": 5.277251243591309, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.028477763757109642, + "learning_rate": 0.0005885028868537, + "loss": 5.295318603515625, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.022988801822066307, + "learning_rate": 0.0005884293161901487, + "loss": 5.249057292938232, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.026115216314792633, + "learning_rate": 0.000588355515561353, + "loss": 5.270636081695557, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.02576332725584507, + "learning_rate": 0.0005882814850267723, + "loss": 5.263096332550049, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.023978352546691895, + "learning_rate": 0.0005882072246460521, + "loss": 5.276605606079102, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.020695513114333153, + "learning_rate": 0.0005881327344790223, + "loss": 5.2141432762146, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01952219195663929, + "learning_rate": 0.0005880580145856986, + "loss": 5.214289665222168, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.022633206099271774, + "learning_rate": 0.0005879830650262813, + "loss": 5.236690998077393, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.023082347586750984, + "learning_rate": 0.0005879078858611557, + "loss": 5.2092156410217285, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.019620463252067566, + "learning_rate": 0.0005878324771508927, + "loss": 5.264451503753662, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.021504493430256844, + "learning_rate": 0.0005877568389562475, + "loss": 5.23444938659668, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.025672364979982376, + "learning_rate": 0.0005876809713381606, + "loss": 5.205289363861084, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.034436631947755814, + "learning_rate": 0.0005876048743577569, + "loss": 5.2047648429870605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.042389530688524246, + "learning_rate": 0.0005875285480763466, + "loss": 5.192008018493652, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.03310716897249222, + "learning_rate": 0.0005874519925554244, + "loss": 5.162734031677246, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.03029288910329342, + "learning_rate": 0.0005873752078566694, + "loss": 5.212151527404785, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04619310051202774, + "learning_rate": 0.0005872981940419459, + "loss": 5.215909004211426, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.05618087574839592, + "learning_rate": 0.0005872209511733027, + "loss": 5.210192680358887, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.03783707693219185, + "learning_rate": 0.0005871434793129726, + "loss": 5.201597690582275, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.03721104562282562, + "learning_rate": 0.0005870657785233735, + "loss": 5.188891410827637, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.033336155116558075, + "learning_rate": 0.0005869878488671075, + "loss": 5.1598615646362305, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.2592363357543945, + "eval_runtime": 57.3343, + "eval_samples_per_second": 42.592, + "eval_steps_per_second": 1.343, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03719741478562355, + "learning_rate": 0.0005869096904069611, + "loss": 5.1646623611450195, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.04888090118765831, + "learning_rate": 0.0005868313032059052, + "loss": 5.171182155609131, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.05073149502277374, + "learning_rate": 0.0005867526873270949, + "loss": 5.183949947357178, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.04187731072306633, + "learning_rate": 0.0005866738428338695, + "loss": 5.176133155822754, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.039445552974939346, + "learning_rate": 0.000586594769789753, + "loss": 5.148100852966309, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.035551708191633224, + "learning_rate": 0.0005865154682584524, + "loss": 5.151512145996094, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.031175289303064346, + "learning_rate": 0.0005864359383038602, + "loss": 5.102468967437744, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.03696789592504501, + "learning_rate": 0.000586356179990052, + "loss": 5.119902610778809, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.04836135357618332, + "learning_rate": 0.0005862761933812875, + "loss": 5.15183162689209, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.05206914246082306, + "learning_rate": 0.0005861959785420106, + "loss": 5.154870986938477, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.046368084847927094, + "learning_rate": 0.000586115535536849, + "loss": 5.130651473999023, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.07146432995796204, + "learning_rate": 0.000586034864430614, + "loss": 5.175804138183594, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.08809789270162582, + "learning_rate": 0.0005859539652883008, + "loss": 5.229716777801514, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.08116360008716583, + "learning_rate": 0.0005858728381750888, + "loss": 5.21910285949707, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.08282570540904999, + "learning_rate": 0.0005857914831563402, + "loss": 5.2336506843566895, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.07403547316789627, + "learning_rate": 0.0005857099002976016, + "loss": 5.251604080200195, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.07024004310369492, + "learning_rate": 0.0005856280896646026, + "loss": 5.267114162445068, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.056330062448978424, + "learning_rate": 0.0005855460513232567, + "loss": 5.201976299285889, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.051763955503702164, + "learning_rate": 0.0005854637853396606, + "loss": 5.1733198165893555, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.039900992065668106, + "learning_rate": 0.0005853812917800945, + "loss": 5.222941875457764, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.04187863692641258, + "learning_rate": 0.0005852985707110221, + "loss": 5.165033340454102, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.042645812034606934, + "learning_rate": 0.0005852156221990901, + "loss": 5.157554626464844, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.03856075555086136, + "learning_rate": 0.0005851324463111289, + "loss": 5.184629440307617, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.029387330636382103, + "learning_rate": 0.0005850490431141516, + "loss": 5.118739128112793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.027584781870245934, + "learning_rate": 0.0005849654126753545, + "loss": 5.100855350494385, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.023655924946069717, + "learning_rate": 0.0005848815550621175, + "loss": 5.154520034790039, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.024163050577044487, + "learning_rate": 0.0005847974703420028, + "loss": 5.105189323425293, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.020369920879602432, + "learning_rate": 0.0005847131585827564, + "loss": 5.113671779632568, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.021030737087130547, + "learning_rate": 0.0005846286198523061, + "loss": 5.0847649574279785, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.01729249767959118, + "learning_rate": 0.0005845438542187638, + "loss": 5.053225517272949, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.018704712390899658, + "learning_rate": 0.0005844588617504236, + "loss": 5.079106330871582, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.02049064077436924, + "learning_rate": 0.0005843736425157621, + "loss": 5.044185161590576, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.020736856386065483, + "learning_rate": 0.000584288196583439, + "loss": 5.041932106018066, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.016543276607990265, + "learning_rate": 0.0005842025240222966, + "loss": 5.032522201538086, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.018414106220006943, + "learning_rate": 0.0005841166249013598, + "loss": 5.037397384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.01611698977649212, + "learning_rate": 0.0005840304992898359, + "loss": 5.038179397583008, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.014310561120510101, + "learning_rate": 0.0005839441472571147, + "loss": 5.0146484375, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.015472214668989182, + "learning_rate": 0.0005838575688727685, + "loss": 5.060323715209961, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.016373401507735252, + "learning_rate": 0.0005837707642065518, + "loss": 5.032100677490234, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.018694665282964706, + "learning_rate": 0.000583683733328402, + "loss": 5.036993026733398, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.019977087154984474, + "learning_rate": 0.0005835964763084378, + "loss": 4.991037368774414, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.023741086944937706, + "learning_rate": 0.000583508993216961, + "loss": 5.043832778930664, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.025061562657356262, + "learning_rate": 0.0005834212841244548, + "loss": 5.016516208648682, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.020848384127020836, + "learning_rate": 0.0005833333491015851, + "loss": 4.986289978027344, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.015086354687809944, + "learning_rate": 0.0005832451882191995, + "loss": 4.9948625564575195, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.015096964314579964, + "learning_rate": 0.0005831568015483274, + "loss": 4.996342658996582, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016932200640439987, + "learning_rate": 0.0005830681891601807, + "loss": 4.998758316040039, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.019936930388212204, + "learning_rate": 0.0005829793511261525, + "loss": 4.959854602813721, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.027658682316541672, + "learning_rate": 0.0005828902875178183, + "loss": 4.968629837036133, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.03299275413155556, + "learning_rate": 0.0005828009984069347, + "loss": 4.994814872741699, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.033717650920152664, + "learning_rate": 0.0005827114838654406, + "loss": 4.976295471191406, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.02969924360513687, + "learning_rate": 0.0005826217439654562, + "loss": 4.969155788421631, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.02846827358007431, + "learning_rate": 0.0005825317787792831, + "loss": 4.989721298217773, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.03670600801706314, + "learning_rate": 0.0005824415883794049, + "loss": 4.988689422607422, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.045570675283670425, + "learning_rate": 0.0005823511728384863, + "loss": 4.974407196044922, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.04564111679792404, + "learning_rate": 0.0005822605322293733, + "loss": 4.966331481933594, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.04173210263252258, + "learning_rate": 0.0005821696666250937, + "loss": 4.925197601318359, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.04228242486715317, + "learning_rate": 0.0005820785760988559, + "loss": 4.965020656585693, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.05144782364368439, + "learning_rate": 0.0005819872607240503, + "loss": 5.0022149085998535, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.049400657415390015, + "learning_rate": 0.0005818957205742478, + "loss": 4.9635114669799805, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.043692708015441895, + "learning_rate": 0.0005818039557232005, + "loss": 4.977723121643066, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.04891742020845413, + "learning_rate": 0.0005817119662448421, + "loss": 4.996312141418457, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.035564981400966644, + "learning_rate": 0.0005816197522132866, + "loss": 4.940932273864746, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04679827764630318, + "learning_rate": 0.0005815273137028292, + "loss": 4.983131408691406, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.05489890277385712, + "learning_rate": 0.0005814346507879459, + "loss": 4.977827548980713, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04640809819102287, + "learning_rate": 0.0005813417635432937, + "loss": 4.954761505126953, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04439452290534973, + "learning_rate": 0.00058124865204371, + "loss": 4.914865970611572, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.04587096720933914, + "learning_rate": 0.0005811553163642131, + "loss": 4.949592590332031, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.044119734317064285, + "learning_rate": 0.0005810617565800019, + "loss": 4.956068992614746, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.043133001774549484, + "learning_rate": 0.0005809679727664559, + "loss": 4.993522644042969, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04315278306603432, + "learning_rate": 0.0005808739649991348, + "loss": 4.894291877746582, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04819472134113312, + "learning_rate": 0.0005807797333537792, + "loss": 4.958306789398193, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.023706436157227, + "eval_runtime": 57.4082, + "eval_samples_per_second": 42.537, + "eval_steps_per_second": 1.341, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.04403403028845787, + "learning_rate": 0.0005806852779063098, + "loss": 4.884971618652344, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.041993748396635056, + "learning_rate": 0.0005805905987328275, + "loss": 4.916353225708008, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.042302004992961884, + "learning_rate": 0.0005804956959096138, + "loss": 4.9012556076049805, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.044500596821308136, + "learning_rate": 0.0005804005695131299, + "loss": 4.915240287780762, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.03914736211299896, + "learning_rate": 0.0005803052196200177, + "loss": 4.885915756225586, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.0361703559756279, + "learning_rate": 0.0005802096463070988, + "loss": 4.829287528991699, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03432302549481392, + "learning_rate": 0.0005801138496513749, + "loss": 4.926834583282471, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.0317804180085659, + "learning_rate": 0.0005800178297300276, + "loss": 4.8707170486450195, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.035261351615190506, + "learning_rate": 0.0005799215866204183, + "loss": 4.897791862487793, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.03276067599654198, + "learning_rate": 0.0005798251204000886, + "loss": 4.864504814147949, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.034757066518068314, + "learning_rate": 0.0005797284311467594, + "loss": 4.890812873840332, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.033726081252098083, + "learning_rate": 0.0005796315189383316, + "loss": 4.831684112548828, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.03074726276099682, + "learning_rate": 0.0005795343838528855, + "loss": 4.8534111976623535, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02215893566608429, + "learning_rate": 0.0005794370259686811, + "loss": 4.8311238288879395, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.02132033370435238, + "learning_rate": 0.0005793394453641581, + "loss": 4.842617034912109, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020825816318392754, + "learning_rate": 0.0005792416421179351, + "loss": 4.84535026550293, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.019334357231855392, + "learning_rate": 0.0005791436163088106, + "loss": 4.839937686920166, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017372334375977516, + "learning_rate": 0.0005790453680157622, + "loss": 4.8115553855896, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.0190406683832407, + "learning_rate": 0.0005789468973179468, + "loss": 4.832542419433594, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.017372574657201767, + "learning_rate": 0.0005788482042947004, + "loss": 4.793632507324219, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.019171632826328278, + "learning_rate": 0.0005787492890255382, + "loss": 4.776987552642822, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.02109278179705143, + "learning_rate": 0.0005786501515901545, + "loss": 4.8216657638549805, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.026234494522213936, + "learning_rate": 0.0005785507920684225, + "loss": 4.8109049797058105, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.0357641875743866, + "learning_rate": 0.0005784512105403944, + "loss": 4.821109771728516, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.03627947345376015, + "learning_rate": 0.000578351407086301, + "loss": 4.810632228851318, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.032253146171569824, + "learning_rate": 0.0005782513817865527, + "loss": 4.7930521965026855, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.03402210772037506, + "learning_rate": 0.0005781511347217375, + "loss": 4.772500991821289, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.0321584977209568, + "learning_rate": 0.000578050665972623, + "loss": 4.781399726867676, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.030242707580327988, + "learning_rate": 0.0005779499756201549, + "loss": 4.802542686462402, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.02361336164176464, + "learning_rate": 0.0005778490637454576, + "loss": 4.771912574768066, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.028781302273273468, + "learning_rate": 0.000577747930429834, + "loss": 4.796924591064453, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.027042683213949203, + "learning_rate": 0.0005776465757547654, + "loss": 4.745218276977539, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.02715984545648098, + "learning_rate": 0.0005775449998019113, + "loss": 4.792944431304932, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.0314096100628376, + "learning_rate": 0.00057744320265311, + "loss": 4.7649993896484375, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.028749193996191025, + "learning_rate": 0.0005773411843903773, + "loss": 4.7980546951293945, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.02655309997498989, + "learning_rate": 0.0005772389450959075, + "loss": 4.746520519256592, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.027497397735714912, + "learning_rate": 0.000577136484852073, + "loss": 4.7490715980529785, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.0343393050134182, + "learning_rate": 0.000577033803741424, + "loss": 4.74436092376709, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.04444412514567375, + "learning_rate": 0.0005769309018466891, + "loss": 4.755973815917969, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.0437556728720665, + "learning_rate": 0.0005768277792507744, + "loss": 4.784368515014648, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0369921550154686, + "learning_rate": 0.0005767244360367638, + "loss": 4.753410339355469, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.05038141459226608, + "learning_rate": 0.0005766208722879192, + "loss": 4.755491256713867, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.04511422663927078, + "learning_rate": 0.00057651708808768, + "loss": 4.760293006896973, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.048762883991003036, + "learning_rate": 0.0005764130835196632, + "loss": 4.755958080291748, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.06749686598777771, + "learning_rate": 0.0005763088586676634, + "loss": 4.802557468414307, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.06451250612735748, + "learning_rate": 0.0005762044136156528, + "loss": 4.812925338745117, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.05436156690120697, + "learning_rate": 0.0005760997484477809, + "loss": 4.807103157043457, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.05233754590153694, + "learning_rate": 0.0005759948632483742, + "loss": 4.782449722290039, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.05519746616482735, + "learning_rate": 0.0005758897581019371, + "loss": 4.808449745178223, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.0578567236661911, + "learning_rate": 0.000575784433093151, + "loss": 4.7676005363464355, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.0588264986872673, + "learning_rate": 0.0005756788883068743, + "loss": 4.805377960205078, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04225790873169899, + "learning_rate": 0.0005755731238281423, + "loss": 4.748775005340576, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.0376368910074234, + "learning_rate": 0.0005754671397421678, + "loss": 4.759791374206543, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03830771893262863, + "learning_rate": 0.0005753609361343402, + "loss": 4.729974746704102, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.03684520348906517, + "learning_rate": 0.0005752545130902256, + "loss": 4.729854583740234, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02983408235013485, + "learning_rate": 0.0005751478706955674, + "loss": 4.78502082824707, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02854904904961586, + "learning_rate": 0.0005750410090362854, + "loss": 4.717472076416016, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.023819150403141975, + "learning_rate": 0.0005749339281984761, + "loss": 4.717124938964844, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.02477158047258854, + "learning_rate": 0.0005748266282684124, + "loss": 4.71781063079834, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.022002534940838814, + "learning_rate": 0.0005747191093325443, + "loss": 4.7480669021606445, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.02133483812212944, + "learning_rate": 0.0005746113714774976, + "loss": 4.693770408630371, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.022407138720154762, + "learning_rate": 0.0005745034147900747, + "loss": 4.720734119415283, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.020850926637649536, + "learning_rate": 0.0005743952393572544, + "loss": 4.745505332946777, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.020420582965016365, + "learning_rate": 0.0005742868452661918, + "loss": 4.688610076904297, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.019014107063412666, + "learning_rate": 0.0005741782326042181, + "loss": 4.673436164855957, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.017393039539456367, + "learning_rate": 0.0005740694014588403, + "loss": 4.6941118240356445, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.017877329140901566, + "learning_rate": 0.0005739603519177419, + "loss": 4.667474269866943, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.017749635502696037, + "learning_rate": 0.0005738510840687821, + "loss": 4.6901092529296875, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.018746376037597656, + "learning_rate": 0.000573741597999996, + "loss": 4.680166721343994, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016260217875242233, + "learning_rate": 0.0005736318937995947, + "loss": 4.6799821853637695, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015099814161658287, + "learning_rate": 0.0005735219715559646, + "loss": 4.680666923522949, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.016539735719561577, + "learning_rate": 0.0005734118313576683, + "loss": 4.639509201049805, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 4.768754482269287, + "eval_runtime": 58.3352, + "eval_samples_per_second": 41.862, + "eval_steps_per_second": 1.32, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.019733518362045288, + "learning_rate": 0.0005733014732934436, + "loss": 4.635312080383301, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.022571036592125893, + "learning_rate": 0.0005731908974522042, + "loss": 4.6041765213012695, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.022238552570343018, + "learning_rate": 0.0005730801039230389, + "loss": 4.648976802825928, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.02115180902183056, + "learning_rate": 0.0005729690927952123, + "loss": 4.623968124389648, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.021789928898215294, + "learning_rate": 0.0005728578641581637, + "loss": 4.620028018951416, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.02593887411057949, + "learning_rate": 0.0005727464181015081, + "loss": 4.591062545776367, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.0347842313349247, + "learning_rate": 0.0005726347547150357, + "loss": 4.632149696350098, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04119519516825676, + "learning_rate": 0.0005725228740887117, + "loss": 4.6254191398620605, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.04895750433206558, + "learning_rate": 0.0005724107763126761, + "loss": 4.621372222900391, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.05752161890268326, + "learning_rate": 0.0005722984614772442, + "loss": 4.61965274810791, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05115870013833046, + "learning_rate": 0.000572185929672906, + "loss": 4.63119649887085, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.05721089243888855, + "learning_rate": 0.0005720731809903263, + "loss": 4.652647018432617, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.0646839588880539, + "learning_rate": 0.000571960215520345, + "loss": 4.642184257507324, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.0682854950428009, + "learning_rate": 0.0005718470333539757, + "loss": 4.676589012145996, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.057402849197387695, + "learning_rate": 0.0005717336345824077, + "loss": 4.6585283279418945, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.05885373800992966, + "learning_rate": 0.0005716200192970043, + "loss": 4.671728134155273, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.05265705659985542, + "learning_rate": 0.0005715061875893032, + "loss": 4.641249656677246, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.05417842045426369, + "learning_rate": 0.0005713921395510166, + "loss": 4.661381721496582, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.06602369248867035, + "learning_rate": 0.0005712778752740307, + "loss": 4.684137344360352, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.06004182621836662, + "learning_rate": 0.0005711633948504066, + "loss": 4.625332832336426, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.04801511764526367, + "learning_rate": 0.0005710486983723787, + "loss": 4.664791107177734, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.04380105435848236, + "learning_rate": 0.0005709337859323561, + "loss": 4.664966583251953, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.03858727216720581, + "learning_rate": 0.0005708186576229218, + "loss": 4.661867618560791, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03240576758980751, + "learning_rate": 0.0005707033135368323, + "loss": 4.5855326652526855, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.028561925515532494, + "learning_rate": 0.0005705877537670184, + "loss": 4.637331008911133, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.026389246806502342, + "learning_rate": 0.0005704719784065846, + "loss": 4.598666667938232, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.02484819106757641, + "learning_rate": 0.0005703559875488088, + "loss": 4.601729393005371, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.02522977627813816, + "learning_rate": 0.0005702397812871429, + "loss": 4.609833717346191, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02589508146047592, + "learning_rate": 0.0005701233597152121, + "loss": 4.574592590332031, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.023374728858470917, + "learning_rate": 0.0005700067229268154, + "loss": 4.61362361907959, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.02465851791203022, + "learning_rate": 0.0005698898710159245, + "loss": 4.639899253845215, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.02258715033531189, + "learning_rate": 0.0005697728040766852, + "loss": 4.5656538009643555, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.018009789288043976, + "learning_rate": 0.0005696555222034162, + "loss": 4.61449670791626, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02121281623840332, + "learning_rate": 0.0005695380254906094, + "loss": 4.553830623626709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.018078701570630074, + "learning_rate": 0.0005694203140329296, + "loss": 4.534579277038574, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.0169303547590971, + "learning_rate": 0.000569302387925215, + "loss": 4.533995628356934, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.016318142414093018, + "learning_rate": 0.0005691842472624764, + "loss": 4.544438362121582, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.016664575785398483, + "learning_rate": 0.0005690658921398977, + "loss": 4.5913286209106445, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.01641504280269146, + "learning_rate": 0.0005689473226528354, + "loss": 4.520743370056152, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.017012735828757286, + "learning_rate": 0.0005688285388968187, + "loss": 4.517453193664551, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01672312617301941, + "learning_rate": 0.0005687095409675499, + "loss": 4.5235915184021, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.016893865540623665, + "learning_rate": 0.000568590328960903, + "loss": 4.543200492858887, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.017530135810375214, + "learning_rate": 0.0005684709029729253, + "loss": 4.54912805557251, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.018698599189519882, + "learning_rate": 0.0005683512630998361, + "loss": 4.522650718688965, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01984071172773838, + "learning_rate": 0.000568231409438027, + "loss": 4.546173095703125, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.023218289017677307, + "learning_rate": 0.0005681113420840619, + "loss": 4.5305094718933105, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.028968270868062973, + "learning_rate": 0.000567991061134677, + "loss": 4.521440505981445, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.032370273023843765, + "learning_rate": 0.0005678705666867805, + "loss": 4.54168701171875, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.033967748284339905, + "learning_rate": 0.0005677498588374524, + "loss": 4.538717269897461, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.028757929801940918, + "learning_rate": 0.0005676289376839452, + "loss": 4.517228603363037, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.029757825657725334, + "learning_rate": 0.0005675078033236827, + "loss": 4.510310173034668, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.03581003099679947, + "learning_rate": 0.0005673864558542605, + "loss": 4.506901264190674, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.03483985736966133, + "learning_rate": 0.0005672648953734462, + "loss": 4.549284934997559, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.031016027554869652, + "learning_rate": 0.000567143121979179, + "loss": 4.530084609985352, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.029373526573181152, + "learning_rate": 0.0005670211357695693, + "loss": 4.492218971252441, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.031661372631788254, + "learning_rate": 0.0005668989368428994, + "loss": 4.502824783325195, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.027251387014985085, + "learning_rate": 0.0005667765252976227, + "loss": 4.49193811416626, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02263176068663597, + "learning_rate": 0.0005666539012323639, + "loss": 4.512019157409668, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.02355228364467621, + "learning_rate": 0.0005665310647459189, + "loss": 4.508917331695557, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.023297371342778206, + "learning_rate": 0.0005664080159372551, + "loss": 4.497771263122559, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.019364114850759506, + "learning_rate": 0.0005662847549055107, + "loss": 4.51470947265625, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.023899901658296585, + "learning_rate": 0.0005661612817499947, + "loss": 4.48771858215332, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.02330983616411686, + "learning_rate": 0.0005660375965701872, + "loss": 4.478496551513672, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.027252234518527985, + "learning_rate": 0.0005659136994657392, + "loss": 4.497672080993652, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.030886363238096237, + "learning_rate": 0.0005657895905364723, + "loss": 4.49789571762085, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02515559457242489, + "learning_rate": 0.0005656652698823788, + "loss": 4.490771293640137, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020107125863432884, + "learning_rate": 0.0005655407376036217, + "loss": 4.481879234313965, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.022076716646552086, + "learning_rate": 0.0005654159938005345, + "loss": 4.494697570800781, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.025525815784931183, + "learning_rate": 0.0005652910385736207, + "loss": 4.517453193664551, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.03005894646048546, + "learning_rate": 0.0005651658720235548, + "loss": 4.453815460205078, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.030178502202033997, + "learning_rate": 0.0005650404942511812, + "loss": 4.493922233581543, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.029249394312500954, + "learning_rate": 0.0005649149053575141, + "loss": 4.453763008117676, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 4.607339382171631, + "eval_runtime": 57.9216, + "eval_samples_per_second": 42.16, + "eval_steps_per_second": 1.329, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.03362950682640076, + "learning_rate": 0.0005647891054437386, + "loss": 4.430700302124023, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.04223684221506119, + "learning_rate": 0.0005646630946112096, + "loss": 4.486915588378906, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.05192171037197113, + "learning_rate": 0.0005645368729614512, + "loss": 4.467541694641113, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.042525868862867355, + "learning_rate": 0.0005644104405961582, + "loss": 4.45252799987793, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.0406370609998703, + "learning_rate": 0.0005642837976171949, + "loss": 4.414548873901367, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.044476233422756195, + "learning_rate": 0.0005641569441265952, + "loss": 4.453492164611816, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.04687326401472092, + "learning_rate": 0.0005640298802265626, + "loss": 4.477725028991699, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.04011611267924309, + "learning_rate": 0.0005639026060194704, + "loss": 4.423356533050537, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.04971884936094284, + "learning_rate": 0.0005637751216078609, + "loss": 4.4853515625, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.05863234028220177, + "learning_rate": 0.0005636474270944461, + "loss": 4.466001033782959, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.057710036635398865, + "learning_rate": 0.0005635195225821072, + "loss": 4.466723442077637, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.05155074968934059, + "learning_rate": 0.0005633914081738945, + "loss": 4.4823198318481445, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.04535634070634842, + "learning_rate": 0.0005632630839730275, + "loss": 4.471219062805176, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.045147854834795, + "learning_rate": 0.0005631345500828946, + "loss": 4.476888179779053, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.03730619326233864, + "learning_rate": 0.0005630058066070533, + "loss": 4.428932189941406, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.034599728882312775, + "learning_rate": 0.0005628768536492299, + "loss": 4.47627592086792, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.032904356718063354, + "learning_rate": 0.0005627476913133193, + "loss": 4.422671794891357, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.03356337174773216, + "learning_rate": 0.0005626183197033856, + "loss": 4.456406593322754, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.035042595118284225, + "learning_rate": 0.0005624887389236609, + "loss": 4.371820449829102, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.03269032761454582, + "learning_rate": 0.0005623589490785462, + "loss": 4.402605056762695, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.028261031955480576, + "learning_rate": 0.0005622289502726108, + "loss": 4.433136940002441, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03301700949668884, + "learning_rate": 0.0005620987426105925, + "loss": 4.388359546661377, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.031664177775382996, + "learning_rate": 0.000561968326197397, + "loss": 4.41859245300293, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.026281699538230896, + "learning_rate": 0.0005618377011380988, + "loss": 4.406128883361816, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.023097911849617958, + "learning_rate": 0.0005617068675379398, + "loss": 4.389717102050781, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.022349286824464798, + "learning_rate": 0.0005615758255023306, + "loss": 4.408173561096191, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.021958226338028908, + "learning_rate": 0.0005614445751368491, + "loss": 4.434830188751221, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024551713839173317, + "learning_rate": 0.0005613131165472415, + "loss": 4.386944770812988, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.028048967942595482, + "learning_rate": 0.0005611814498394216, + "loss": 4.402279376983643, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02733907848596573, + "learning_rate": 0.0005610495751194708, + "loss": 4.403069496154785, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.020858559757471085, + "learning_rate": 0.0005609174924936384, + "loss": 4.406920909881592, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.021491149440407753, + "learning_rate": 0.0005607852020683409, + "loss": 4.385799407958984, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.021495329216122627, + "learning_rate": 0.0005606527039501621, + "loss": 4.4252495765686035, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.019708355888724327, + "learning_rate": 0.0005605199982458535, + "loss": 4.388864994049072, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.01886274851858616, + "learning_rate": 0.0005603870850623338, + "loss": 4.335962295532227, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.01955341547727585, + "learning_rate": 0.0005602539645066884, + "loss": 4.373560905456543, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.018962012603878975, + "learning_rate": 0.0005601206366861706, + "loss": 4.380022048950195, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.016663335263729095, + "learning_rate": 0.0005599871017081999, + "loss": 4.421034336090088, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01676873490214348, + "learning_rate": 0.0005598533596803631, + "loss": 4.382775783538818, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.014379126951098442, + "learning_rate": 0.0005597194107104137, + "loss": 4.41373872756958, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.015897583216428757, + "learning_rate": 0.000559585254906272, + "loss": 4.377284049987793, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.016513120383024216, + "learning_rate": 0.000559450892376025, + "loss": 4.41200065612793, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.016625918447971344, + "learning_rate": 0.000559316323227926, + "loss": 4.381187438964844, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.016028188169002533, + "learning_rate": 0.000559181547570395, + "loss": 4.353909015655518, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01682914048433304, + "learning_rate": 0.0005590465655120183, + "loss": 4.382326126098633, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.017078889533877373, + "learning_rate": 0.0005589113771615486, + "loss": 4.334949493408203, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.018182938918471336, + "learning_rate": 0.0005587759826279046, + "loss": 4.348130226135254, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01940886303782463, + "learning_rate": 0.0005586403820201713, + "loss": 4.339404106140137, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.020644692704081535, + "learning_rate": 0.0005585045754475995, + "loss": 4.358404159545898, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02433430217206478, + "learning_rate": 0.0005583685630196064, + "loss": 4.3660502433776855, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.027824537828564644, + "learning_rate": 0.0005582323448457744, + "loss": 4.34050178527832, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.02973582223057747, + "learning_rate": 0.0005580959210358523, + "loss": 4.348322868347168, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.030396128073334694, + "learning_rate": 0.0005579592916997542, + "loss": 4.356224060058594, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.029760530218482018, + "learning_rate": 0.0005578224569475599, + "loss": 4.357470989227295, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.02855013869702816, + "learning_rate": 0.0005576854168895147, + "loss": 4.335353374481201, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03376635164022446, + "learning_rate": 0.0005575481716360292, + "loss": 4.331794738769531, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.034431129693984985, + "learning_rate": 0.0005574107212976796, + "loss": 4.336880207061768, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.03524135425686836, + "learning_rate": 0.000557273065985207, + "loss": 4.31537389755249, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.03152058273553848, + "learning_rate": 0.0005571352058095179, + "loss": 4.37662410736084, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02878640592098236, + "learning_rate": 0.0005569971408816838, + "loss": 4.292388916015625, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.029663654044270515, + "learning_rate": 0.0005568588713129409, + "loss": 4.344346523284912, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.027895759791135788, + "learning_rate": 0.0005567203972146906, + "loss": 4.365072727203369, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.026465769857168198, + "learning_rate": 0.000556581718698499, + "loss": 4.334959506988525, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.03191852569580078, + "learning_rate": 0.000556442835876097, + "loss": 4.361069679260254, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.03535444661974907, + "learning_rate": 0.0005563037488593799, + "loss": 4.349765777587891, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03451581299304962, + "learning_rate": 0.0005561644577604074, + "loss": 4.343142509460449, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029810845851898193, + "learning_rate": 0.000556024962691404, + "loss": 4.307485580444336, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.027523217722773552, + "learning_rate": 0.0005558852637647584, + "loss": 4.366406440734863, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03043208085000515, + "learning_rate": 0.0005557453610930235, + "loss": 4.356929779052734, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.03058922104537487, + "learning_rate": 0.0005556052547889161, + "loss": 4.348942279815674, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.032052166759967804, + "learning_rate": 0.0005554649449653177, + "loss": 4.3496809005737305, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.03170694038271904, + "learning_rate": 0.0005553244317352731, + "loss": 4.315959930419922, + "step": 720 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.069613361332224e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r90-baseline/checkpoint-720/training_args.bin b/runs/l2r90-baseline/checkpoint-720/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9 --- /dev/null +++ b/runs/l2r90-baseline/checkpoint-720/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85 +size 4856